My videoclip for the fabulous and enchanting Melbourne duo, Time For Dreams.
by j p, October 11, 2017 0 comments
Software Adventure-Time! Newest* video kid on the block = Mitti, a “modern, feature-packed but easy-to-use pro video cue playback solution for events, theatre, audiovisual shows, performances and exhibitions,” – coming from the same stable that brought us Vezer (the timeline based midi/osc/dmx sequencer) and COGE (the versatile VJ software). (By *Newest, I mean it’s been around since late 2016, but since then, Mitti has enjoyed a steady rate of notable additions and updates.)
After all the work that can go into video material for an event, playback control can sometimes be left as an afterthought – it’s not unknown to see videos being played back from video editing software and ‘live-scrubbed’, or to watch users flipping between their desktop and powerpoint / keynote / Quicktime etc. VJ software of course brings a flexibility and reliability to playback control – taking care of the basics such as fading to black, or looping upon finish of clip, cross-fading, or simply – avoiding the desktop suddenly appearing on the main projected screen. The flexibility of most VJ software is also one of it’s limitations – the strength of real-time effects and mixing, tending to make interfaces more obtuse than they need to be for users seeking simple playback. And when getting simple playback exactly right becomes important, for events, theatre, installations etc – this is where Mitti seems to be aiming at, in the ballpark of apps like Playback Pro, QLAB or perhaps Millumin, where cues and critical timing are given more priority than visual effects.
Running at it’s most minimal, with a playlist of clips, Mitti appears deceptively simple – but comes dense with custom controls at every level.
The Interface – a strength would seem to be the efforts spent in making a very clear and intuitive interface for Mitti – it comes across as clean and easy to navigate, with extended options available where they might be expected. (And for further depth – go to the control menu, choose Mitti, then Preferences, to see a very well organised array of options)
Timing – Exact timing control can be critical, and Mitti boasts low latency, a GPU playback engine, and can run from an SMPTE based internal clock, or slave to external hardware or software sources. If you know what these Mitti capacities mean, you’re possibly the intended audience: external MTC (MIDI Timecode), LTC (Linear Timecode), SMPTE offsetting, Jam-Sync.
Cues – You can create and trigger cues for videos, images, cameras (and native Blackmagic support), syphon and NDI sources – each with nuanced options, and easily add or adjust in/out points per clip. Cues can be set to loop, and the playlist can be paused between cues, until ready to start the next cue.
Nuanced control of fades, transitions – eg individual control per cue over fade in or outs, and over 30 ISF based video transition options
Remote Controllability + Sync – Hello MMC (Midi Machine Control), MSC (MIDI show control), hardware and software midi controllers, OSC + Art-Net (DMX over ethernet). Multiple versions of Mitti on the same network can be easily synced.
Output finessing – Includes colour controls, video effects, audio channel routing and for multiple video displays – Mitti provides individual 4-corner warping for each, and edge blending between overlapping projectors. Mitti can also output syphon, blackmagic and NDI.
Below, showing the various cue preference options:
Video output options:
Project Preferences, which include detailed options for each item on the left:
This can be important when dealing with unusual projects or computer quirks… and Imimot boast great support – “Our average first response time was only 3 hours 5 minutes in the past 7 days!”, as well as extensive FAQ / tips and support documentation:
A mac computer – running 10.10-10.12.
$299US for perpetual licence (for 2 computers) (30% edu discount available)
$79US rental licence for 30 days from first activation.
Solid, reliable software that’ll be of interest to anyone involved with running video for time-critical events, theatre and installations. Double thumbs up!
Tamas, below in Imimot HQ, was nice enough to answer some questions about why he made Mitti….
With the ecosystem of video software that exists – what inspired you to add MITTI to it?
The idea of creating Mitti is coming from Vezér feature requests – the funny thing with this is Vezér was also born by a couple of CoGe feature requests 🙂 A lot of Vezér users were searching …. for an any-to-use but remote controllable video playback solution, which plays nice with Vezér, or even requested video playback functionality in Vezér. Adding video playback functions to Vezér does not sound reasonable for me, the app was not designed that way, and I wanted to leave it as a “signal processes”/show control tool instead of redesigning the whole app. After doing some research I’ve found there were no app that time on the market I could offer to Vezér users which is easy to setup and use, lightweight, and controllable by various protocols that Vezér supports. So I’ve started to create one. The original plan was to make something really basic, but once I’ve started to speak about the project to my pals and acquaintances, I’ve realised there is a need for an easy to use app on the market with pro features, and by pro features I mean timecode sync, multi-output handling, capture card support.
What are some contexts you imagine MITTI being used?
Mitti is targeting the presentation, theatre, broadcast and exhibitions market, usually where reliable cue-based media playback is needed – and this is where Mitti’s current user base come from: event producer companies, theatre, visual techs of touring artists, composers working together with DAWs, etc.
What interests you about NDI?
I believe NDI is the next big thing after Syphon. Now you can share frames between computers even running different operating systems without or minimal latency, using cheap network gear or already exists network infrastructure. And there are even hardwares coming with native NDI support!
What are some interesting ways you’ve enabled Vezer+Mitti interplay?
The coolest thing is NMC I believe – our in-house network protocol which enabled sending transport and locate command thru the network. Vezér sends NMC commands by defaults and Mitti listens to them by default, which means you can make Mitti scrub on its timeline and receive start/stop commands from Vezér with no setup.
An other big thing is the OSC Query. This is not strictly Mitti and Vezér related. OSC Query is a protocol – still in draft mode yet – proposed by mrRay from Vidvox to discover an OSC-enabled app’s OSC address space. As far as I know only Mitti and Vezér supporting this protocol on the market, but hopefully others will join pretty soon, since this going to be a game changer in my opinion.
You can even copy-paste Cues from Mitti to Vezér.
Why is Mitti priced more expensive than COGE?
This is a rather complex topic, but basically Mitti has been designed to a fairly different market than CoGe. Also CoGe is highly underpriced in my opinion – well, pricing things is far more complex stuff than I imagined when CoGe hit the prime time – but that is a whole different topic.
(See also : COGE review and interview )
by j p, June 28, 2017 0 comments
I spent a few nights in a hospital basement last year, projecting video and controlling lights for The General Assembly – onto a room filled with paper strips, while audiences roamed between rooms for mini-sets. It was part of Melbourne Music Week and super fun – the video below shows it up nicely.
Will be doing projections for TGA again this saturday at The Toff In Town :
by j p, February 2, 2017 0 comments
Melbourne, as the most Nathan Barley of Australian cities, so easily lampooned for its population of bushranger bearded baristas with half-baked app ideas, makes a strong argument for being Australia’s Portland. Perfectly placed then, for reviewing Lumen – new real-time visual software coded by Jason Grlicky in downtown Portland, which tries to add some contemporary twists to the quirky history of video synthesis.
A mac based app (needing OSX 10.8 or later) for ‘creating engaging visuals in real-time’… with a ‘semi-modular design that is both playable and deep.. the perfect way to get into video synthesis.’ In other words – it’s a software based video synthesiser, with all the noodling, head-scratching experiments and moments of delightful serendipity this implies. A visual synthesiser – that can build up images from scratch, then rhythmically modify and refine them over time. It has been thoughtfully put together though, so despite the range of possibilities – it’s also very quickly ‘playable’ – and always suggesting there’s plennnnttttyyyy of room to explore.
While the underlying principles of hardware based video synthesisers are being milked here to good effect – a lot of the merits of Lumen are in the ways they’ve managed to make these principles easily accessible with well considered interface design. It has been divided into 3 sections – a preset browser (which also features a lovely X/Y pad for interpolating between various presets), a knob panel interface, and a patch panel interface. It’s a very skeuomorphic design, but it also cleverly takes the software to places where hardware couldn’t go (more on that later).
What should be evident in those screengrabs, is that experimentation is easy- and there’s a lot of depth to explore. The extensive reference material helps a lot with the latter. And as you can see, they can’t help but organise that beautifully on their site:
Lumen comes pre-loaded with 150+ presets, so it’s immediately satisfying upon launch, to be able to jump between patches and see what kind of scope and visual flavours are possible.
… and it’s easy to copy and remix presets, or export and swap them – eg on the Lumen slack channel.
Although all are planned, only midi exists in Lumen so far, but it’s beautifully integrated. With a midi controller (or a phone/tablet app sending OSC to a midi translating app on your computer) – Lumen really comes into it’s own, and the real-time responsiveness can be admired. Once various parameters are connected via midi control, those of course can effectively be made to be audioreactive, by sending signals from audioreactively controlled parameters in other software. Native integration will be nice when it arrives though.
Decent syphon integration of course opens a whole range of possibilities…. Lumen’s output can be easily piped into software like VDMX or COGE for use as a graphic source or texture, or mapping software like madmapper. At the moment there are some limitations with aspect ratios and output sizes, but that’s apparently being resolved in a near-future update.
With the ability to import video via syphon though, Lumen can reasonably considered as an external visual effects unit. Lumen can also take in camera feeds for processing, but it’s the ability to take in a custom video feed that can make it versatile – eg video clips created for certain visual ideas, or the output of a composition in a mapping program.
This screengrab below shows the signal going into Lumen from VDMX, and also out of lumen back into VDMX. Obviously, at some point this inevitably means feedback, and all the associated fun/horror.
Also worth a look: Lumen integration with VDMX – post by artist Wiley Wiggins – and accompanying video tutorial (including a nice top to use osculator as a means of enabling touchosc finger-gliding control over the square pad interpolation of Lumen snapshots).
There’s an army of lovers of abstracted visuals that are going to auto-love Lumen, but it has scope too for others looking for interesting ways to add visual textures, and play with real-time visual effects on video feeds. It could feasibly have an interesting place in a non-real-time video production pipeline too. Hopefully in a few years, we’ll be awash in a variety of real-time visual synthesis apps, but for now Lumen is a delightfully designed addition to the real-time video ecosystem.
– What inspired you to develop Lumen?
I’ve always loved synthesizers, but for most of my life that was limited to audio synths. As soon as I’d heard about video synthesis, I knew I had to try it for myself! The concept of performing with a true video instrument – one that encourages real-time improvisation and exploration – really appeals to me.
Unfortunately, video synths can be really expensive, so I couldn’t get my hands on one. Despite not being able to dive in (or probably because of it), my mind wouldn’t let it go. After a couple failed prototypes, one morning about I woke up with a technical idea for how I could emulate the analog video synthesis process in software. At that point, I knew that my path was set…
– When replicating analogue processes within software – what have been some limitations / happy surprises?
There have been so many happy accidents along the way. Each week during Lumen’s development, I discovered new techniques that I didn’t think would be possible with the instrument. There are several presets that I included which involve a slit-scan effect that only works because of the specific way I implemented feedback, for instance! My jaw dropped when I accidentally stumbled on that. I can’t wait to see what people discover next.
My favorite part about the process is that the laws of physics are just suggestions. Software gives me the freedom to deviate from the hardware way of doing things in order to make it as easy as possible for users. The way that Lumen handles oscillator sync is a great example of this.
Can you describe a bit more about that freedom to deviate from hardware – in how Lumen handles oscillator sync?
In a traditional video synth oscillator, you’ll see the option to sync either to the line rate or to the vertical refresh rate, which allows you to create vertical or horizontal non-moving lines. When making Lumen, I wanted to keep the feeling of control as smooth as possible, so I made oscillator sync a knob instead of a switch. As you turn it clockwise, the scrolling lines created by the oscillator slow down, then stop, then rotate to create static vertical lines. It’s a little thing, but ultimately allows for more versatile output and more seamless live performance than has ever been possible using hardware video synths.
Were there any other hardware limitations that you were eager to exploit the absence of within software?
At every turn I was looking for ways to push beyond what hardware allows without losing the spirit of the workflow. The built-in patch browser is probably the number-one example. Being able to instantly recall any synth settings allows you to experiment faster than with a hardware synth, and having a preset library makes it easier to use advanced patching techniques.
The Snapshots XY- Pad, Undo & Redo, and the Transform/K-Scope effects are all other examples of where we took Lumen beyond what hardware can do today. Honestly, I think we’re just scratching the surface with what a software video instrument can be.
How has syphon influenced software development for you?
I had an epiphany a couple years back where I took a much more holistic view of audio equipment. After using modular synths for long enough, I realized that on a certain level, the separation between individual pieces of studio equipment is totally artificial. Each different sound source, running through effects, processed in the mixer – all of that is just part of a larger system that works together to create a part of a song. This thinking led me to create my first app, Polymer, which is all about combining multiple synths in order to play them as a single instrument.
For me, Syphon and Spout represent the exact same modular philosophy – the freedom to blend the lines between individual video tools and to treat them as part of a larger system. Being able to tap into that larger system allowed me to create a really focused video instrument instead of having to make it do everything under the sun. Thanks to technologies like Syphon, the future of video tools is a very bright place!
What are some fun Lumen + Syphon workflows you enjoy – or enjoy seeing users play with?
My favorite workflow involves setting up Syphon feedback loops. You just send Lumen’s output to another VJ app like CoGe or VDMX, put some effects on it, then use that app’s output as a camera input in Lumen. It makes for some really unpredictable and delightful results, and that’s just from the simplest possible feedback loop!
What are some things you’re excited about on the Lumen roadmap ahead?
We have so many plans for things to add and refine. I’m particularly excited about improving the ways that Lumen connects with the outside world – be that via new video input types, control protocols, or interactions with other programs. We’re working on adding audio-reactivity right now, which is going to be a really fun when it ships. Just based on what we’ve seen in development so far, I expect it to add a whole new dimension to Lumen while keeping the workflow intuitive. It’s a difficult balance to strike, but that’s our mission – never to lose sight of the immediacy of control while adding new features.
by j p, November 30, 2016 0 comments
I recently animated some vintage botanical illustrations for an interactive exhibition installation at The Royal Botanic Gardens, Sydney. It was fun to collaborate with Robert Jarvis ( zeal.co ) on this – who programmed the interactivity (incorporating childrens’ webcam photos into the various creatures and plant-life storylines), as well as with D.A. Calf ( dacalf.com ) who brought the world to life so well. And a special shout-out to Luke Dearnley and Sophie Daniel who produced it.
by j p, October 24, 2016 0 comments