My videoclip for the fabulous and enchanting Melbourne duo, Time For Dreams.
by j p, October 11, 2017 0 comments
Software Adventure-Time! Newest* video kid on the block = Mitti, a “modern, feature-packed but easy-to-use pro video cue playback solution for events, theatre, audiovisual shows, performances and exhibitions,” – coming from the same stable that brought us Vezer (the timeline based midi/osc/dmx sequencer) and COGE (the versatile VJ software). (By *Newest, I mean it’s been around since late 2016, but since then, Mitti has enjoyed a steady rate of notable additions and updates.)
After all the work that can go into video material for an event, playback control can sometimes be left as an afterthought – it’s not unknown to see videos being played back from video editing software and ‘live-scrubbed’, or to watch users flipping between their desktop and powerpoint / keynote / Quicktime etc. VJ software of course brings a flexibility and reliability to playback control – taking care of the basics such as fading to black, or looping upon finish of clip, cross-fading, or simply – avoiding the desktop suddenly appearing on the main projected screen. The flexibility of most VJ software is also one of it’s limitations – the strength of real-time effects and mixing, tending to make interfaces more obtuse than they need to be for users seeking simple playback. And when getting simple playback exactly right becomes important, for events, theatre, installations etc – this is where Mitti seems to be aiming at, in the ballpark of apps like Playback Pro, QLAB or perhaps Millumin, where cues and critical timing are given more priority than visual effects.
Running at it’s most minimal, with a playlist of clips, Mitti appears deceptively simple – but comes dense with custom controls at every level.
The Interface – a strength would seem to be the efforts spent in making a very clear and intuitive interface for Mitti – it comes across as clean and easy to navigate, with extended options available where they might be expected. (And for further depth – go to the control menu, choose Mitti, then Preferences, to see a very well organised array of options)
Timing – Exact timing control can be critical, and Mitti boasts low latency, a GPU playback engine, and can run from an SMPTE based internal clock, or slave to external hardware or software sources. If you know what these Mitti capacities mean, you’re possibly the intended audience: external MTC (MIDI Timecode), LTC (Linear Timecode), SMPTE offsetting, Jam-Sync.
Cues – You can create and trigger cues for videos, images, cameras (and native Blackmagic support), syphon and NDI sources – each with nuanced options, and easily add or adjust in/out points per clip. Cues can be set to loop, and the playlist can be paused between cues, until ready to start the next cue.
Nuanced control of fades, transitions – eg individual control per cue over fade in or outs, and over 30 ISF based video transition options
Remote Controllability + Sync – Hello MMC (Midi Machine Control), MSC (MIDI show control), hardware and software midi controllers, OSC + Art-Net (DMX over ethernet). Multiple versions of Mitti on the same network can be easily synced.
Output finessing – Includes colour controls, video effects, audio channel routing and for multiple video displays – Mitti provides individual 4-corner warping for each, and edge blending between overlapping projectors. Mitti can also output syphon, blackmagic and NDI.
Below, showing the various cue preference options:
Video output options:
Project Preferences, which include detailed options for each item on the left:
This can be important when dealing with unusual projects or computer quirks… and Imimot boast great support – “Our average first response time was only 3 hours 5 minutes in the past 7 days!”, as well as extensive FAQ / tips and support documentation:
A mac computer – running 10.10-10.12.
$299US for perpetual licence (for 2 computers) (30% edu discount available)
$79US rental licence for 30 days from first activation.
Solid, reliable software that’ll be of interest to anyone involved with running video for time-critical events, theatre and installations. Double thumbs up!
Tamas, below in Imimot HQ, was nice enough to answer some questions about why he made Mitti….
With the ecosystem of video software that exists – what inspired you to add MITTI to it?
The idea of creating Mitti is coming from Vezér feature requests – the funny thing with this is Vezér was also born by a couple of CoGe feature requests 🙂 A lot of Vezér users were searching …. for an any-to-use but remote controllable video playback solution, which plays nice with Vezér, or even requested video playback functionality in Vezér. Adding video playback functions to Vezér does not sound reasonable for me, the app was not designed that way, and I wanted to leave it as a “signal processes”/show control tool instead of redesigning the whole app. After doing some research I’ve found there were no app that time on the market I could offer to Vezér users which is easy to setup and use, lightweight, and controllable by various protocols that Vezér supports. So I’ve started to create one. The original plan was to make something really basic, but once I’ve started to speak about the project to my pals and acquaintances, I’ve realised there is a need for an easy to use app on the market with pro features, and by pro features I mean timecode sync, multi-output handling, capture card support.
What are some contexts you imagine MITTI being used?
Mitti is targeting the presentation, theatre, broadcast and exhibitions market, usually where reliable cue-based media playback is needed – and this is where Mitti’s current user base come from: event producer companies, theatre, visual techs of touring artists, composers working together with DAWs, etc.
What interests you about NDI?
I believe NDI is the next big thing after Syphon. Now you can share frames between computers even running different operating systems without or minimal latency, using cheap network gear or already exists network infrastructure. And there are even hardwares coming with native NDI support!
What are some interesting ways you’ve enabled Vezer+Mitti interplay?
The coolest thing is NMC I believe – our in-house network protocol which enabled sending transport and locate command thru the network. Vezér sends NMC commands by defaults and Mitti listens to them by default, which means you can make Mitti scrub on its timeline and receive start/stop commands from Vezér with no setup.
An other big thing is the OSC Query. This is not strictly Mitti and Vezér related. OSC Query is a protocol – still in draft mode yet – proposed by mrRay from Vidvox to discover an OSC-enabled app’s OSC address space. As far as I know only Mitti and Vezér supporting this protocol on the market, but hopefully others will join pretty soon, since this going to be a game changer in my opinion.
You can even copy-paste Cues from Mitti to Vezér.
Why is Mitti priced more expensive than COGE?
This is a rather complex topic, but basically Mitti has been designed to a fairly different market than CoGe. Also CoGe is highly underpriced in my opinion – well, pricing things is far more complex stuff than I imagined when CoGe hit the prime time – but that is a whole different topic.
(See also : COGE review and interview )
by j p, June 28, 2017 0 comments
I spent a few nights in a hospital basement last year, projecting video and controlling lights for The General Assembly – onto a room filled with paper strips, while audiences roamed between rooms for mini-sets. It was part of Melbourne Music Week and super fun – the video below shows it up nicely.
Will be doing projections for TGA again this saturday at The Toff In Town :
by j p, February 2, 2017 0 comments
Melbourne, as the most Nathan Barley of Australian cities, so easily lampooned for its population of bushranger bearded baristas with half-baked app ideas, makes a strong argument for being Australia’s Portland. Perfectly placed then, for reviewing Lumen – new real-time visual software coded by Jason Grlicky in downtown Portland, which tries to add some contemporary twists to the quirky history of video synthesis.
A mac based app (needing OSX 10.8 or later) for ‘creating engaging visuals in real-time’… with a ‘semi-modular design that is both playable and deep.. the perfect way to get into video synthesis.’ In other words – it’s a software based video synthesiser, with all the noodling, head-scratching experiments and moments of delightful serendipity this implies. A visual synthesiser – that can build up images from scratch, then rhythmically modify and refine them over time. It has been thoughtfully put together though, so despite the range of possibilities – it’s also very quickly ‘playable’ – and always suggesting there’s plennnnttttyyyy of room to explore.
While the underlying principles of hardware based video synthesisers are being milked here to good effect – a lot of the merits of Lumen are in the ways they’ve managed to make these principles easily accessible with well considered interface design. It has been divided into 3 sections – a preset browser (which also features a lovely X/Y pad for interpolating between various presets), a knob panel interface, and a patch panel interface. It’s a very skeuomorphic design, but it also cleverly takes the software to places where hardware couldn’t go (more on that later).
What should be evident in those screengrabs, is that experimentation is easy- and there’s a lot of depth to explore. The extensive reference material helps a lot with the latter. And as you can see, they can’t help but organise that beautifully on their site:
Lumen comes pre-loaded with 150+ presets, so it’s immediately satisfying upon launch, to be able to jump between patches and see what kind of scope and visual flavours are possible.
… and it’s easy to copy and remix presets, or export and swap them – eg on the Lumen slack channel.
Although all are planned, only midi exists in Lumen so far, but it’s beautifully integrated. With a midi controller (or a phone/tablet app sending OSC to a midi translating app on your computer) – Lumen really comes into it’s own, and the real-time responsiveness can be admired. Once various parameters are connected via midi control, those of course can effectively be made to be audioreactive, by sending signals from audioreactively controlled parameters in other software. Native integration will be nice when it arrives though.
Decent syphon integration of course opens a whole range of possibilities…. Lumen’s output can be easily piped into software like VDMX or COGE for use as a graphic source or texture, or mapping software like madmapper. At the moment there are some limitations with aspect ratios and output sizes, but that’s apparently being resolved in a near-future update.
With the ability to import video via syphon though, Lumen can reasonably considered as an external visual effects unit. Lumen can also take in camera feeds for processing, but it’s the ability to take in a custom video feed that can make it versatile – eg video clips created for certain visual ideas, or the output of a composition in a mapping program.
This screengrab below shows the signal going into Lumen from VDMX, and also out of lumen back into VDMX. Obviously, at some point this inevitably means feedback, and all the associated fun/horror.
Also worth a look: Lumen integration with VDMX – post by artist Wiley Wiggins – and accompanying video tutorial (including a nice top to use osculator as a means of enabling touchosc finger-gliding control over the square pad interpolation of Lumen snapshots).
There’s an army of lovers of abstracted visuals that are going to auto-love Lumen, but it has scope too for others looking for interesting ways to add visual textures, and play with real-time visual effects on video feeds. It could feasibly have an interesting place in a non-real-time video production pipeline too. Hopefully in a few years, we’ll be awash in a variety of real-time visual synthesis apps, but for now Lumen is a delightfully designed addition to the real-time video ecosystem.
– What inspired you to develop Lumen?
I’ve always loved synthesizers, but for most of my life that was limited to audio synths. As soon as I’d heard about video synthesis, I knew I had to try it for myself! The concept of performing with a true video instrument – one that encourages real-time improvisation and exploration – really appeals to me.
Unfortunately, video synths can be really expensive, so I couldn’t get my hands on one. Despite not being able to dive in (or probably because of it), my mind wouldn’t let it go. After a couple failed prototypes, one morning about I woke up with a technical idea for how I could emulate the analog video synthesis process in software. At that point, I knew that my path was set…
– When replicating analogue processes within software – what have been some limitations / happy surprises?
There have been so many happy accidents along the way. Each week during Lumen’s development, I discovered new techniques that I didn’t think would be possible with the instrument. There are several presets that I included which involve a slit-scan effect that only works because of the specific way I implemented feedback, for instance! My jaw dropped when I accidentally stumbled on that. I can’t wait to see what people discover next.
My favorite part about the process is that the laws of physics are just suggestions. Software gives me the freedom to deviate from the hardware way of doing things in order to make it as easy as possible for users. The way that Lumen handles oscillator sync is a great example of this.
Can you describe a bit more about that freedom to deviate from hardware – in how Lumen handles oscillator sync?
In a traditional video synth oscillator, you’ll see the option to sync either to the line rate or to the vertical refresh rate, which allows you to create vertical or horizontal non-moving lines. When making Lumen, I wanted to keep the feeling of control as smooth as possible, so I made oscillator sync a knob instead of a switch. As you turn it clockwise, the scrolling lines created by the oscillator slow down, then stop, then rotate to create static vertical lines. It’s a little thing, but ultimately allows for more versatile output and more seamless live performance than has ever been possible using hardware video synths.
Were there any other hardware limitations that you were eager to exploit the absence of within software?
At every turn I was looking for ways to push beyond what hardware allows without losing the spirit of the workflow. The built-in patch browser is probably the number-one example. Being able to instantly recall any synth settings allows you to experiment faster than with a hardware synth, and having a preset library makes it easier to use advanced patching techniques.
The Snapshots XY- Pad, Undo & Redo, and the Transform/K-Scope effects are all other examples of where we took Lumen beyond what hardware can do today. Honestly, I think we’re just scratching the surface with what a software video instrument can be.
How has syphon influenced software development for you?
I had an epiphany a couple years back where I took a much more holistic view of audio equipment. After using modular synths for long enough, I realized that on a certain level, the separation between individual pieces of studio equipment is totally artificial. Each different sound source, running through effects, processed in the mixer – all of that is just part of a larger system that works together to create a part of a song. This thinking led me to create my first app, Polymer, which is all about combining multiple synths in order to play them as a single instrument.
For me, Syphon and Spout represent the exact same modular philosophy – the freedom to blend the lines between individual video tools and to treat them as part of a larger system. Being able to tap into that larger system allowed me to create a really focused video instrument instead of having to make it do everything under the sun. Thanks to technologies like Syphon, the future of video tools is a very bright place!
What are some fun Lumen + Syphon workflows you enjoy – or enjoy seeing users play with?
My favorite workflow involves setting up Syphon feedback loops. You just send Lumen’s output to another VJ app like CoGe or VDMX, put some effects on it, then use that app’s output as a camera input in Lumen. It makes for some really unpredictable and delightful results, and that’s just from the simplest possible feedback loop!
What are some things you’re excited about on the Lumen roadmap ahead?
We have so many plans for things to add and refine. I’m particularly excited about improving the ways that Lumen connects with the outside world – be that via new video input types, control protocols, or interactions with other programs. We’re working on adding audio-reactivity right now, which is going to be a really fun when it ships. Just based on what we’ve seen in development so far, I expect it to add a whole new dimension to Lumen while keeping the workflow intuitive. It’s a difficult balance to strike, but that’s our mission – never to lose sight of the immediacy of control while adding new features.
by j p, November 30, 2016 0 comments
I recently animated some vintage botanical illustrations for an interactive exhibition installation at The Royal Botanic Gardens, Sydney. It was fun to collaborate with Robert Jarvis ( zeal.co ) on this – who programmed the interactivity (incorporating childrens’ webcam photos into the various creatures and plant-life storylines), as well as with D.A. Calf ( dacalf.com ) who brought the world to life so well. And a special shout-out to Luke Dearnley and Sophie Daniel who produced it.
by j p, October 24, 2016 0 comments
One of the video-art greats passed away recently – RIP Bill Etra, who leaves behind a huge legacy for his work at the intersections of art and technology. Below, Bill Etra demonstrates the functions of the Rutt/Etra Video Synthesizer. (1974)
“Bill Etra, an artist and inventor who, with a partner, created a video animation system in the early 1970s that helped make videotape a more protean and accessible medium for many avant-garde artists, died on Aug. 26 near his home in the Bronx. He was 69.
The cause was heart failure, said his wife, Rozalyn Rouse Etra. Mr. Etra had spinal stenosis for many years and was mostly bedridden when he died.
Mr. Etra and Steve Rutt created the Rutt/Etra video synthesizer, an analog device studded with knobs and dials that let a user mold video footage in real time and helped make video a more expressive art form. Among the artists who used it were Nam June Paik, regarded by many as the father of video art, and Woody and Steina Vasulka, who founded the Kitchen performance space in downtown Manhattan in 1971.”
“The dream was to create a compositional tool that would allow you to prepare visuals like a composer composes music,” Mr. Etra wrote. “I called it then and I call it now the ‘visual piano,’ because with the piano the composer can compose an entire symphony and be sure of what it will sound like. It was my belief then, and it is my belief now after 40 years of working towards this, that this will bring about a great change and great upwelling of creative work once it is accomplished.”
“Developed in 1972, the RUTT/ETRA Video Synthesizer was one of the first commercially available computerized video animation systems. It employed proprietary analog computer technology to perform real time three dimensional processing of the video image. In the first use of computer animation in a major Hollywood picture, Steve Rutt, working directly with Sidney Lumet, used the Rutt/Etra to create the animated graphic for the film’s “UBS” Television Network.”
Rutt-Etra-Izer is a WebGL emulation of the classic Rutt-Etra video synthesizer, by Felix Turner, which ‘replicates the Z-displacement, scanned-line look of the original, but does not attempt to replicate it’s full feature set’. The demo allows you to drag and drop your own images, manipulate them and save the output. Images are generated by scanning the pixels of the input image from top to bottom, with scan-line separated by the ‘Line Separation’ amount. For each line generated, the z-position of the vertices is dependent on the brightness of the pixels.
Realtime audio-reactive processing application by Felix Turner. It scans input image pixels and draws scanlines with Z position proportional to the brightness of the pixel. Audio level modifies z-pos from top to bottom.
by j p, September 19, 2016 0 comments
Hermitude Concert Visuals from jeanpoole on Vimeo.
Am glad to finally upload that edit-medley – because creating a set of concert visuals for Hermitude was one of my favourite projects last year, seeing it from drawing-board and sketch paper, through to the stage screen. Hermitude had approached (having worked together on Dr.Seuss Meets Elefant Traks at Sydney’s Graphic Festival in 2012) – about developing video for their tour promoting Dark Night, Sweet Light – and wanted a visual set that suited their music, would work well within a hectic stage lighting environment, and was diverse but felt like a coherent, consistent show.
To suit Hermitude’s fun and festive sound and their dynamic live performances – I developed an overall visual style palette to enhance that, and mapped out a visual choreography for the show. And though I was excited about making some Hermitude clips of my own, it was also an exciting opportunity to collaborate with some talented animators, coders and cinematographers. It was fantastic to be able to work with these artists to craft the Hermitude set:
Neil Sanders – a Melbourne hand-drawn illustrator and animator extraordinaire, famous for his signature organic tumblr loops…
Ori Toor – another hand-drawn abstraction loop specialist, beaming pixels to us from the Middle East.
Colin E. White – moodily stylised New York animator.
Brad Hammond – A Melbourne 3D Unity animation ninja + coder. (And shout-out to Kejiro Takahashi from Japan, for his ongoing publishing of Unity software addons… )
Stu Gibson – A Tasmanian surf + aerial cinematographer, who was very generous with his wild coastline footage (which I used to make the Bermuda Bay clip below)
It was also a pleasure to develop this visual set over time, because Luke ‘Dubs’ + Angus ‘El Gusto’ (aka Hermitude) are so down to earth and friendly, despite their relentless touring and acclaim, as are the whole Elefant Traks crew – especially their tireless manager (and collaborator) Urthboy and Luke Dearnley (Sub Bass Snarl), their wizardly tour manager (who designed a clever + efficient video rig featuring live cams – for routing and controlling their stage video feeds).
A lot of pixel-sweat across quite a few months, but …
.. so satisfying to see it all come together in the end.
by j p, March 18, 2016 0 comments
Music video made for the Cumbia Cosmonauts Mexico Remixes EP, based on footage shot during our tour of Mexico in Dec 2014.
(Shout outs to all the fine people we met in Guadalajara, Monterrey, Metepec, Tijuana, Mexico City and Oaxaca!)
The video and EP launched on CassetteBlog in Jan 2016.
by j p, January 27, 2016 1 Comment
by j p, November 6, 2015 0 comments
I was lucky enough recently to catch a film-talk panel between director Joshua Oppenheimer and John Safran, at the Melbourne International Film festival. Having just seen the Look of Silence earlier that day, and already in awe of the brave and audacious film-making from the earlier companion film (The Act of Killing) – it was humbling and a privilege to hear about some of what went into the making of the film – and what some of its’ impacts have been since.
Given that Indonesia has not officially or publically discussed the mass killings that happened in 1965-66 (supposedly to get rid of a communist threat) – and that many of the perpetrators are entrenched in power today, it’s quite remarkable that these two films got made – prompted national discussions about them – and that the second film was given official recognition:
“On November 10, 2014, 2,000 people came to the official and public premiere of the film in Jakarta, and on December 10, 2014 – International Human Rights Day – there were 480 public screenings of the film across Indonesia. The screenings of the film in Indonesia has been sponsored by the National Human Rights Commission of Indonesia and the Jakarta Arts Council.” ( Via wikipedia)
Incredibly, after the first film – which featured the ‘surreal / defensive(?)’ boasting of one of the mass-killers – an Indonesian journalist saw the film, and persuaded their magazine to send out investigative journalists to document similar people in 60 different locations across Indonesia – and then published all of these in one go – alongside an in depth reaction to Oppenheimer’s film – which broke the silence, and allowed Indonesian media to move past the taboo of discussing these events.
Regardless of your awareness of this Indonesian mass killing, these are powerful films on many levels – well worth hunting down.
The film focuses on the perpetrators of the Indonesian killings of 1965–66 in the present day; ostensibly towards the communist community where almost a million people were killed.
Invited by Oppenheimer, Anwar recounts his experiences killing for the cameras, and makes scenes depicting their memories and feelings about the killings. The scenes are produced in the style of their favorite films: gangster,western, and musical.
The name “Anonymous” appears 49 times under 27 different crew positions in the credits. These crew members still fear revenge from the death-squad killers.
When the government of Indonesia was overthrown by the military in 1965, Anwar and his friends were promoted from small-time gangsters who sold movie theatre tickets on the black market to death squad leaders. They helped the army kill more than one million alleged communists, ethnic Chinese, and intellectuals in less than a year. As the executioner for the most notorious death squad in his city, Anwar himself killed hundreds of people with his own hands. Today, Anwar is revered as a founding father of a right-wing paramilitary organization that grew out of the death squads. The organization is so powerful that its leaders include government ministers, and they are happy to boast about everything from corruption and election rigging to acts of genocide.
The Act of Killing is about killers who have won, and the sort of society they have built.
In The Act of Killing, Anwar and his friends agree to tell us the story of the killings. But their idea of being in a movie is not to provide testimony for a documentary: they want to star in the kind of films they most love from their days scalping tickets at the cinemas. We seize this opportunity to expose how a regime that was founded on crimes against humanity, yet has never been held accountable, would project itself into history.
And so we challenge Anwar and his friends to develop fiction scenes about their experience of the killings, adapted to their favorite film genres – gangster, western, musical. They write the scripts. They play themselves. And they play their victims.
via Wikipedia entry for Indonesian killings of 1965–66
“the CIA denies active involvement in the killings, though it was later revealed that the American government provided extensive lists of Communists to Indonesian death squads.”
The Act of Killing Trailer:
“Through Oppenheimer’s footage of perpetrators of the 1965 Indonesian genocide, a family of survivors discovers how their son was murdered, as well as the identities of the killers. The documentary focuses on the youngest son, an optometrist named Adi, who decides to break the suffocating spell of submission and terror by doing something unimaginable in a society where the murderers remain in power: he confronts the men who killed his brother and, while testing their eyesight, asks them to accept responsibility for their actions. This unprecedented film initiates and bears witness to the collapse of fifty years of silence.”
The Look of Silence Trailer:
by j p, August 12, 2015 0 comments
Teaser video for an upcoming videoclip featuring the visual highlights of the late 2014 Cumbia Cosmonauts tour of Mexico.
The video will help launch the Mexico Remixes EP (coming soon via cassetteblog.com), which will include remixes by Bumb Ay, Yelram Selectah, Malacopa Bros, HNRY + Sonido Satanas – and The Ghost whose remix of Cumbia Sampuesana you can hear above.
Full video coming soon…
by j p, June 23, 2015 0 comments
I was in Hobart to do triple-Screen Video Projections at ‘ Faux Mo‘, which is the afterparty venue each night for the MOFO Festival, connected to the MONA gallery in Hobart, Tasmania. It tends to be eclectic – here’s the program.
Highlights included doing video for – the Frikstailers, (rad Argentinian duo), ZA! (Barcelona crazies!), Chupacabra (Hobart based Latin American band),Vulgargrad (Russian-gypsy-punk!), The Osaka Ramones (Aka Shonen Knife doing Ramones covers), Mad Professor (dubb-b-b), and The Bird– and getting to animate the illustrations of Hobart artists – Tom O’ Hern and Dane Chisolm.
(click the image below to see a larger version)
Other festival good-times while wandering:
– some of Tim Hecker‘s set had me transported, Zammuto (ex-Books) have a pretty engaging and fun audiovisual live set, playing on Robin Fox’s laser-organ was surprisingly engaging, Filastine + Nova delivered a very tightly crafted audiovisual set, there was some impromptu whisky-bar parkour by Jamin, Keith Deverell‘s beautifully shot film installation used mirrors and a walk-through screen very effectively, and finally managed a few moments inside the inflatable labyrinth by Architects of Air, before departing. Thanks Hobart!
Click through for a photo-slideshow below, or browse directly.
by j p, January 22, 2015 0 comments
Am super excited – it’ll be my first time in any of the Americas. From Nov 26 – Dec 22 I’ll be wandering through Mexico City, Oaxaca, Tijuana, as well as Cuernavaca, Metepec and a few other places in between.
[[UPDATE >> Would you like me to send a MEXICAN POSTCARD to your letterbox? ]]
I’ll be doing live-video for the Cumbia Cosmonauts tour (you can help the Cumbia Cosmonauts tour crowdfunding campaign below, + get assorted Mexican goods in the mail… postcards, marketplace mixtapes, tshirts etc) – and outside of that I hope to be holidaying through Mexican landscapes, crazy butterfly reserves, Aztec ruins, mexican wrestling matches, visiting unusual rooftops, haggling about weird vintage Mexican Masks, practicing my duolingo Espańol (si, level one!)….
…and where possible, doing some live video projection + filming outdoors with any Mexican artists / film-makers / performers.
Love to hear from anybody with Mexico-tips, or weird filming locations / fun people to film / project with etc
And aye, flying across the Pacific Ocean to Mexico means going via Los Angeles – so it seemed insane not to be spending my first few days in the United States on the way back.
Love to hear any Los Angeles tips – places to visit/ good rooftop filming locations / people interested in outdoor projections – or Death Valley trips?
And here’s my low-budget Cumbia Cosmonauts video-clip from back in 2012:
by j p, November 14, 2014 0 comments
Chancha Via Circuito – a favourite listen in recent years – has a new album out – Amansara (Wonderwheel Recordings). I first discovered his enchanting atmospheres and mixing on his wonderful ZZK Records mixtape (promoting his previous album Rio Arriba). His music seems to thrive best in mixtapes (see also Mixtape Cumbiero European Tour 2013 and a mixtape at Testpressing for new album ), reminding at times of early Future Sound of London and their wandering from soundscape to rhythm and back again. There’s a warmth to this music, and despite a slower tempo, there’s a momentum to it all as well. Recommendo!
Oh and a special shout-out too, for Paula Duro, who makes the enchanting artwork for Chancha (and featured in the backlayer of the collage above), as well as much of her own cool stuff. Check out her playful cosmic palette at flickr.
by j p, November 12, 2014 0 comments