Remote Band Rehearsals
Share using http://tinyurl.com/RemoteBandRehearsals
The information below is no longer being actively maintained, but may still be useful as a starting point and to guide the reader to other valuable online information resources.
Last updated October 5, 2024.
If you’re on a computer (not a mobile device) and don’t see the Table of Contents on the left, click the box icon to display it. On a mobile device, you’ll just have to scroll down.
This document is long - here’s some links to help you get started:
I’m a relative beginner keyboard player, with a basic understanding of music theory, chord structure, etc. I play in a live-performance cover band (The Real Imposters) with my wife, and we’re both also choral singers. During the 2020 COVID-19 pandemic, all musical groups had to discontinue in-person rehearsals, and I sought out solutions for remote band rehearsals, some of which are also applicable for remote choral rehearsals.
This document compiles resources that I’ve found and what I’ve learned about enabling remote band rehearsals via the internet. Although primarily focused on groups of 6 or fewer people, the information can be applied to larger groups, such as school bands or orchestras, choruses and choirs, etc. - I’ve seen groups of 30+ musicians/singers performing “live” together remotely. Any additions/corrections can be sent to Chris Rimple at chrisrimple@yahoo.com.
“Latency” is the delay from when one person makes a sound (such as playing an instrument or singing) until another person hears it. Latency is measured in milliseconds (ms), which is thousandths of a second. The human brain can perceive a delay of 15-20ms or more, and it can become difficult for musicians to stay in sync when latency exceeds 50ms. 1ms of latency is roughly equal to 1 foot of distance (for sound traveling through the air), so 25ms of latency is the same as trying to play an instrument in sync with someone that is 25 feet away. Because the speed of light is much faster than the speed of sound, large orchestras, choral groups, and marching bands (all with performers spread over a large performance area) use a “visual metronome” to stay in time - often this is a director/conductor. Smaller groups, like a band on stage, convert their sound to digital format, which can be transmitted over cables (at slightly less than the speed of light, but much faster than the speed of sound) to monitors/amplifiers for the other band members to hear.
When two people are separated by hundreds or thousands of miles and the sound is transmitted via the internet, the latency can become quite large. As a result, latency can make it very difficult for two people using a standard video conferencing application (Zoom, Skype, Microsoft Teams, etc.) to play or sing together - they’ll never be able to stay in sync. Applications designed to enable remote band rehearsals overcome the problem of latency in two ways:
Swingology band has a short blog post with basic information on a number of the Real Time and Delayed Sync applications.
Wikipedia has a page comparing the technical features of Real Time and Delayed Sync applications.
I’m using Real Time applications for remote band rehearsals, specifically Jamulus, which has proven to be the most successful Real Time application and has an excellent support community. Jamulus is becoming a popular choice for musicians and my band has been using it successfully for remote rehearsals - I’ve also performed on the Jamulus WorldJam. When connected to servers near my home, latency is less than 40ms. Choral groups are also turning to Jamulus, with a number of servers for singers now available, and even large choruses using it for rehearsals.
I’ve also spent some hours using Delayed Sync applications, primarily Jamtaba. I’ve created a bunch of loops to play over (as backing tracks), and I’ve done some jamming and improvising with musicians around the world. I’ve had the opportunity to hear and join great musicians playing R&B, Jazz, Rock, and a variety of electronic genres. When I can’t find a server that interests me, I join an empty server, start up a loop, begin playing, and wait for others to join.
For band rehearsals, I use my full equipment stack and wired internet connection with a Real Time application - I sometimes do the same with a Delayed Sync application. When I just want to jam and keep it simple, I use my Mac on WiFi, 25-key keytar with MIDI adapter, MainStage for sound generation, and BlackHole to route sound to a Delayed Sync application - it’s a very portable solution that I can use from anywhere in my house, as well as other locations, since Delayed Sync applications are less susceptible to WiFi latency.
Regardless of whether you’re using a Real Time or Delayed Sync application, the equipment requirements are identical.
Very few of the remote band rehearsal solutions listed here are available as mobile applications. Most applications have versions available for Windows, Mac, and Linux. Check the individual system requirements to determine if your computer is sufficient. As a general rule, any computer with a processor speed of 2.0GHz or faster and 4GB of RAM should work.
Important Note: Disable all background applications that might impact your computer’s performance. That includes disabling WiFi, Bluetooth, cloud sync (Google Drive, Microsoft OneDrive, etc.), file backup (including Time Machine for Mac and File History for Windows), and any other applications that may be running in the background. On Mac, many background applications show an icon on the right side of the Menu Bar - on Windows, they often appear as System Tray icons on the right side of the Taskbar.
My equipment
You want the fastest internet connection you can get, for both download and upload. Some applications may say that 500kbps both ways is sufficient, but I recommend nothing less than 1mb down and 1mb up - you’ll be far better off with 5mb up or greater. Most home “high speed” internet is at least 10mb down and 1mb up, so you’re likely good if you’ve got cable or fiber internet service. You can test your internet speed at SpeedTest.net.
You also want the most “stable” internet connection you can get, one with very little packet loss and jitter. For most people, fiber provides a far more stable connection than cable, so consider fiber if you have that option. Note that some fiber connections are not “to the house”, but are only “to the street”, and you want fiber all the way to the modem in your house, or you may not see all the benefits.
All of the applications recommend a wired connection, rather than WiFi. In my own testing, I’ve found that Wifi can work in some circumstances, but I don’t recommend it - although I’m getting 300mb+ down and 40mb up on WiFi, I still get better results when wired - WiFi is subject to much more packet loss and jitter. If your computer is normally only connected via WiFi, you’ll probably need a long Ethernet cable and maybe an adapter. WiFi can work for Delayed Sync applications, since they’re less dependent on keeping latency low.
You’ll want to plug the Ethernet cable directly into your cable/fiber modem, unless you’ve disabled Network Address Translation (NAT) and are using your own router - in that case, plug the Ethernet cable into your router.
Important Note: In addition to a wired connection, it’s important to disable WiFi for all other devices during a remote band rehearsal, which can help to reduce jitter. Disable WiFi at the router or disconnect as many devices as possible. If you need WiFi enabled for a second device (such as when connected to Zoom for the Jamulus WorldJam), rename your WiFi SSID so that your other devices don’t automatically connect to it, then return it to the regular name when done.
My equipment
Most guitars, basses, keyboards, and microphones designed for live performance are analog instruments. If it plugs into an amplifier, or is amplified using analog microphones (such as acoustic drums), it’s an analog instrument. Usually, analog instruments provide output via 1/4-inch “instrument” or XLR cables.
Headsets and earbuds designed for use with computers and phones may be analog instruments. They usually provide input/output via 1/8-inch cables. If you only want to sing (no other instruments) and aren’t too concerned about quality, an inexpensive “gaming” headset or medium-quality earbuds are sufficient, and no Audio Interface will be needed.
Some instruments, such as keyboards, may offer both analog and digital output. If you’re using an Audio Interface, you’ll have to decide whether to route the analog output to the Audio Interface or the digital output directly to the computer (via USB). For simplicity, I recommend using the analog output.
My equipment
Digital instruments and microphones are designed to output directly to a computer or other digital receiver, usually via a USB cable. Some of these instruments may not have a speaker or any analog output, instead functioning as a “controller” for music applications - examples include keyboards sending MIDI to applications like MainStage or GarageBand, a digital drum kit that sends MIDI to a rack-mounted synthesizer, and so on. Any microphone that connects to a computer via a USB cable is a digital microphone.
Headsets designed for use with computers may be digital instruments. They usually provide input/output via a USB cable. If you only want to sing (no other instruments) and aren’t too concerned about quality, an inexpensive “gaming” headset is sufficient, and no Audio Interface will be needed.
As noted in Analog Instruments & Microphones, I recommend using analog output whenever possible, for simplicity. For example, my Alesis Vortex 2 keytar is a digital-only controller - it generates no sound and has no speaker or analog output. But, instead of sending MIDI Out to a music application like MainStage to generate digital sound, I send MIDI Out from the keytar to my Nord Electro 6D MIDI In, and the Nord generates analog sound which then goes via 1/4-inch Line Out to my Audio Interface.
If you want to use a digital instrument, you’ll need to route the output from it to your Audio Interface (usually through your computer) or an application. More details are provided in Digital Audio Workstation. To route the output to or from an application, you may need to use a Virtual Audio Driver for audio routing “within” your computer.
My equipment
You should not listen to yourself play or sing “live” - instead, you listen to yourself played back as others hear you (which is slightly delayed), helping to ensure synchronization. To achieve this, you need wired headphones, earbuds, or similar that are connected to your Audio Interface (or possibly your computer if you’re doing any pre-/post-processing of sound before/after outputting to a remote band rehearsal application).
In a quiet environment, earbuds are sufficient. In a louder environment, you’ll want to block out background noise with headphones. “Noise cancelling” is not necessary and not particularly helpful, unless background noise is relatively constant and high pitched (such as roadway noise). If you will be singing, you may prefer “open back” headphones that allow your own voice to reach your ears “live”, or block it completely with “closed” headphones.
You should not be monitoring your “live” sound and should not be listening through speakers, particularly if you have a microphone (since the microphone will likely pick up sound from the speakers, creating a feedback loop and echo).
Important Note: Headphones/earbuds must be wired, not wireless. Bluetooth headphones/earbuds cannot be used because they add significant latency.
Important Note: For some remote band rehearsal applications, the headphones/earbuds must be capable of outputting 48KHz - some can only output lower bitrates, or have a range of output options but exclude 48KHz.
My equipment
An Audio Interface (AI) converts audio signals from analog-to-digital and digital-to-analog. The AI connects to a computer and one or more analog devices (guitar, bass, keyboard, microphone, etc.) - anything played on the device is transmitted in digital form to the computer, where it can be recorded or passed on to a remote band rehearsal application. As well, anything played on the computer is transmitted to the AI, where it can be output to analog amplifiers and other analog devices.
Important Note: For some remote band rehearsal applications, the AI must be capable of outputting 48KHz - some AIs can only output 44KHz, or have a range of output options but exclude 48KHz.
Example AIs
My equipment
A Digital Audio Workstation (DAW) is an application for creating and editing digital multi-track recordings. DAWs can be used with Digital Instruments & Microphones, or Analog Instruments & Microphones connected via an AI. A DAW will only be necessary if you are using one or more Digital Instruments & Microphones or need pre-/post-processing of sound before/after outputting to a remote band rehearsal application.
Common DAWs include Logic Pro, MainStage, GarageBand, Audacity, Reaper, Pro Tools, Cakewalk, Cubase, and others. Note that Audacity (Windows, Mac) , Reaper (Windows, Mac), and GarageBand (Mac) are free - others may be as well.
Important Note: For some remote band rehearsal applications, the DAW must be capable of outputting 48KHz - some DAWs can only output 44KHz, or have a range of output options but exclude 48KHz. GarageBand cannot output 48KHz.
My equipment
A virtual audio driver may be needed with some band rehearsal applications if you want to (1) use more inputs than the application allows or (2) route audio from one application to another, such as when you want to use Jamulus or Jamtaba with a DAW but without an AI. For example, if you want to use Jamulus (which is limited to 2 inputs) with an AI that has 4 devices plugged in (keys, drums, 2 mics), you’ll need to route the audio from the AI (4 inputs) to a DAW (4 inputs) to a virtual audio driver (2 outputs) to Jamulus (2 inputs).
For Mac, virtual audio drivers include Soundflower (MacOS 10.10 Yosemite and earlier), BlackHole (MacOS 10.11 El Capitan and later, including Catalina), LoopBack, and others. If you’re a non-technical Mac user, you may find LoopBack easiest to use, although it’s not free like some. For Windows, there’s ASIO4ALL, JACK, VoiceMeeter, and others. Reaper (DAW) for Windows includes an audio routing solution (ReaRoute) and is free.
For a little background on audio routing for Mac, see the “Using a DAW” section of Jim Amsden’s blog post and the support page for BlackHole. For Windows, see the “Soundcard Settings” section of Brian Mountford’s Google doc. For any platform, also see the “Routing The DAW To And From Jamulus” section of Brian Mountford’s Google doc, which provides an excellent (and simple) overview of merging inputs and pre-/post-processing of audio before sending to, or when received from, a remote band rehearsal application.
Important Note: For some remote band rehearsal applications, the virtual audio driver must be capable of outputting 48KHz - some virtual audio drivers can only output 44KHz, or have a range of output options but exclude 48KHz.
My Equipment
For some remote band rehearsal applications, any headphones/earbuds, AI, DAW, and virtual audio driver used must be capable of inputting and outputting 48KHz.
To check and set the bitrate on a Mac:
Checking and setting the output bitrate on Windows depends on the device driver being used. For ASIO4ALL:
From some remote band rehearsal applications, security “permission” may be required to access microphones, instruments, AI, etc.
To check and set access permissions on a Mac:
To check and set access permissions on Windows:
Real Time applications are potentially good for playing through an entire song end-to-end and as a replacement for in-person band rehearsals. These applications solve latency by decreasing it as much as possible, usually through a combination of short physical distance between participants, high speed internet connections, and reducing/removing elements that can increase latency.
Most Real Time applications use a hub-and-spoke model, requiring all users to connect to the same server. As each user plays, the sound they generate is transmitted to the server, then sent back to all users (including the original). In the hub-and-spoke model, it’s critical that the latency between users and the server be as low as possible. Other Real Time applications use a peer-to-peer model, requiring all users to connect to all other users. As each user plays, the sound they generate is transmitted to all other users. In the peer-to-peer model, it’s critical that the latency between all users be as low as possible. Of the two, the hub-and-spoke model is more forgiving for users with lower speed computers and lower speed internet connections.
Jamulus uses the hub-and-spoke model and can be very successful when all users are within 100 miles of the server. When looking for servers, Jamulus provides latency values, allowing the user to select the nearest/fastest server for the best experience. Jamulus has public servers, but it’s also reasonably simple to set up a private server, including running in a virtual machine on Microsoft Azure, Google Cloud, or Amazon Web Services. My band has used Jamulus, on a server within 10 miles and at an overall latency of 30-40ms, and has had no meaningful issues staying in sync.
Jamulus is limited to 2 mono input channels, which can restrict users with multiple instruments and microphones. That limitation can be overcome by running multiple instances of the Jamulus application or by merging multiple inputs to a single input, which is then sent to Jamulus - this is usually accomplished with a Virtual Audio Driver or DAW.
The Jamulus WorldJam community of musicians has developed around Jamulus, including the hosting of WorldJam Nights, a regular, live audio/video broadcast of musicians playing from around the world.
General Information & FAQ
Installation
Instructional Videos
I haven’t watched all these yet and need to order them, highlight important contents, etc.
Usage Videos
Server Installation
Server Instructional Videos
Social Media & Forums
A number of musicians in Western and Northern Europe came together to create the Jamulus WorldJam (WJ), a weekly “jam session” that is broadcast live via Facebook and YouTube, using Jamulus for audio and Zoom for video. Although called a “jam”, WJ is a structured event with hosts, sound checks, and a predefined set list, but remains open to everyone and participants are not professionals.
As awareness and popularity of the WJ has grown, participation has expanded worldwide, but latency is still a limiter. Some European and USA musicians have attempted to play together, but latency often exceeds 100ms and makes staying in sync exceptionally difficult. WJ offers multiple servers with the results fed into the WJ - for example, a server in Seattle WA (USA) is used by musicians in nearby states and Canadian provinces to play together on the WJ, while servers in Iowa and Virginia are used by musicians near those states.
Songs can be added to an upcoming set list by anyone, then it’s up to that musician to recruit others to play the needed parts. Usually that’s accomplished by suggesting the song in the appropriate Facebook group, where other musicians will learn about it and volunteer. In general, the WJ observes a “maximum 2 songs per lead singer per week” rule, but no limit on instrument participation, and the organizers are mindful to give space to new participants and those that didn’t participate during the prior week. Rehearsing isn’t a requirement, but it’s common for musicians on a song to try to rehearse a few times before the WJ - if the participants don’t feel comfortable that a song is ready, it can always be rescheduled to a later date.
The set list is locked the day before the WJ, and includes sound check, show schedule, and other participant information. It’s published to the Jamulus WorldJam Facebook group.
When it’s time for the WJ, first-time participants join a sound and video check server 1.5 hours before showtime, then join a waiting room server at least 15 minutes before showtime. The waiting rooms have a “virtual monitor”, so musicians can hear the WJ while waiting to perform, and remain connected to the Zoom video conference but don’t send video until it’s time for their song(s). Some musicians run Zoom on the same computer with Jamulus, while others use a second device. If using a second device, it’s best if the device is not using WiFi.
The WJ is simultaneously broadcast on Facebook (720p) and YouTube (1080p). Audio is not meaningfully altered from what’s sent by the individual musicians - a limiter catches peaks, light compression reduces dynamic range, and a slight equalizer brings out a few frequencies. While Jamulus allows individual musicians (listening to the “virtual monitor”) to control the balance that they hear, such as “more vocals and less guitar”, the broadcasts to Facebook and YouTube are mixed live by a WJ volunteer sound engineer. Musicians on the same song can improve their balance by doing their sound checks together, ensuring a better relative balance before performing.
The WJ prefers to receive a separate Jamulus feed for each instrument or voice, which can be accomplished by running multiple Jamulus instances if needed. If a musician has only 2 inputs, such as voice and guitar, it’s OK to run a single Jamulus instance with inputs split L/R, Mono output, and Pan centered.
Important Notes if you want to participate in the WJ...
General Information & FAQ
Social Media & Forums
A number of choral groups are using Jamulus for rehearsals and “drop-in singing”, most notably barbershop choruses and quartets, which have begun hosting Jamulus servers with “barbershop” in the name. The organizers of the Jamulus WorldJam also hosted the first Jamulus Choral Festival, which featured up to 19 singers at one time. By default, Jamulus servers are limited to 10 users, but some have been configured for 25 or 50 - experiments have shown that Jamulus struggles with more than 20 users, so it’s not yet a solution for larger choruses. But some tests have been done to “chain” multiple Jamulus servers, possibly allowing as many as 180 users at one time.
If you’re a singer wanting to use Jamulus, much of the Equipment section can be ignored. At a minimum, you will need:
Configure your computer’s Sound settings to use the headset/earbuds for Input (Microphone) and Output (Speakers), then run Jamulus and select “in:System/out:System” for your Device. Connect to a server and speak, and you should hear your voice returned from the server. If other users are on the server, you’ll hear them and they should be able to hear you - if so, you’re ready to sing.
If you decide that you want a higher-quality microphone, there are a number of reasonable-quality “gaming” headsets available, as well as the Blue Yeti line of USB microphones. Check with other Jamulus singers to see what they’re using.
You may also want to review these instructional links:
Important Note: When connecting to Jamulus servers, the List dropdown at the top of the server list defaults to “All Genres”, but most Jamulus servers being used by choral groups will be found under “Genre Classical/Folk/Choral”, so try other choices if you don’t see the server you’re seeking.
Social Media & Forums
How I’m using Jamulus.
How I have Jamulus configured for my full equipment stack when using 2 Jamulus instances.
When my band is learning a new song, we’ll sometimes play along with the original recording, which can often be found on YouTube or a streaming music service (iTunes, Spotify). I route the audio to Jamulus by running an additional instance of Jamulus.
My employer uses Microsoft Teams, a web conferencing and text chat application. For a recent event, some co-workers and I rehearsed a few songs via Jamulus, then performed for the company via a Teams conference. I routed the Jamulus output to Teams, but did not attempt to route the Teams output back to Jamulus, since that could have created a feedback loop. Instead, I ran the Teams application on a separate device, with the microphone and speaker muted, then unmuted the speaker between songs so that I could hear the audience's questions and comments.
How I have Jamulus configured for my full equipment stack when using 1 Jamulus instance and merging more than 2 inputs via a DAW and Virtual Audio Driver. I need to review and update this section.
How I’m using Jamulus servers.
How I have Jamulus Server configured, when I choose to run a local server.
When I want to record band rehearsals, I start Jamulus Server from a Terminal command line
How I have configured my Jamulus cloud server.
My server is usually public, but when I want it to be private to prevent possible interruptions, I edit the jamulus.service file to remove the directory server parameter:
Prior to my band’s rehearsals, I reboot the server:
When my band is rehearsing and I want to record, I edit the jamulus.service file to add the recording parameter to the end:
Lately, I’ve been leaving recording enabled on the server at all times:
To update the server to a new version of Jamulus:
To downgrade the server to an older version of Jamulus, replace “r3_8_0” with the desired version:
How I have configured my Jamulus cloud server.
My server is usually public, but when I want it to be private to prevent possible interruptions, I edit the jamulus.service file to remove the central server parameter - see my Debian configuration above for examples.
Prior to my band’s rehearsals, I reboot the server - see my Debian configuration above for examples.
When my band is rehearsing and I want to record, I don’t have to change anything, because recording is always enabled (see ExecStart parameters above). Because recording is enabled at all times, I’ve enabled a cron job, running at 12pm daily (server local time), to delete recordings that are more than 24 hours old - see my Debian configuration above for examples.
To update the server to a new version of Jamulus:
With recording enabled on my Jamulus server, and having administrative access, I download recordings using Cyberduck (although any application supporting SFTP download will work). How I’ve configured Cyberduck:
After a Jamulus session has completed, I run Cyberduck and connect to that Bookmark. Cyberduck displays a list of list recording folders - I select the most recent and Download.
When the download is complete, I open the recording using Audacity, where I can trim the beginning and end, save specific sections of the Jamulus session, save individual performers, and so on.
JamKazam is similar to Jamulus. Some people have said that it used to be good, but now the servers are overloaded and it crashes regularly. It doesn’t show latency values when looking for a server to join. I’ve spent a few hours trying it, but prefer Jamulus.
General Information & FAQ
Installation
Instructional Videos
Usage Videos
Social Media & Forums
SoundJack uses the peer-to-peer model. I haven’t tested it yet, but the sample videos seem promising.
General Information & FAQ
Installation
Instructional Videos
Usage Videos
Social Media & Forums
JackTrip uses the peer-to-peer model. I haven’t tested it yet, but setup and configuration is all done via command line, so not a good choice for anyone that isn’t significantly tech skilled. It seems to be good for 2 musicians, but may be difficult to get working with 3 or more. Some people have reported that it crashes frequently.
General Information & FAQ
Installation
Instructional Videos
Usage Videos
Social Media & Forums
Delayed Sync applications are great for unstructured jamming, “noodling” on an instrument, and learning to improvise, but not good for playing through an entire song end-to-end or as a replacement for in-person band/chorus rehearsals. These applications solve latency by increasing it, from hundredths of a second to multiple measures. To quote the Ninjam documentation:
NINJAM provides a solution by making latency much longer. Latency in NINJAM is measured in measures... The NINJAM client records and streams synchronized intervals of music between participants. Just as the interval finishes recording, it begins playing on everyone else’s client. So when you play through an interval, you’re playing along with the previous interval of everybody else, and they’re playing along with your previous interval.
Think of it like this: the drummer counts in and plays 8 measures of a song, then everyone else hears the count in and plays along - meanwhile, the drummer is always 8 measures ahead and hears everyone else as 8 measures behind. Yes, the drummer could repeat the first 8 measures to be “in sync” with everyone else, but then it gets worse: all band members are playing the first 8 measures together, but then they hear what everyone played on measure 1 when they are ready to play measure 9.
Where Delayed Sync applications can be valuable is when playing with a short, looping backing track or playing the verse-chorus portion of a song repeatedly. For example, the song “Ghostbusters” has a repetitive verse and chorus pattern of 8m intro, 16m verse, 16m chorus - playing that pattern repeatedly on drums and bass can make it easy for other instruments to improvise “on top”, but they won’t be able to play the full song from beginning to end.
Ninjam is not an application, but rather a set of servers that are accessed using an application like Reaper, Jamtaba, and others. Ninbot is a subset of the public servers, run by a group of Ninjam fans. I’ve used Ninjam a few times and enjoy it, but recommend the Jamtaba user interface for beginners, since Reaper is a DAW with much more capability and complexity.
General Information & FAQ
Installation
Instructional Videos
Usage Videos
Server Installation
Server Installation Videos
Social Media & Forums
Jamtaba provides a simple interface to Ninjam servers, including a map showing the location of users accessing each server, which can help in finding other musicians that speak the same language. It also provides the ability to “listen in” to a server before joining it, to see if the genre/style of music being played is of interest to you, or just to hear what others are doing with Ninjam.
Jamtaba is limited to 2 input channels with a sub-channel for each - if you need more inputs, you’ll have to use a DAW to merge those inputs to a single output before sending it to Jamtaba.
General Information & FAQ
Installation
Instructional Videos
Usage Videos
Social Media & Forums
How I have Jamtaba configured for my full equipment stack.
How I have Jamtaba configured for my MIDI keytar, which requires a DAW for sound generation.
Seems to be a solution similar to Ninjam / Reaper, but I haven’t tested it yet.
Using GarageBand, I’ve created a variety of loops for use with Ninjam/Reaper/Jamtaba, that could also be used with other applications. I’ve shared the GarageBand files, allowing anyone to download and edit them, as well as MP3s. The filename for each loop includes the BPM, length in measures, key signature, and chord progression, as well as a loop name. Drum-only loop filenames omit the key signature and chord progression.
In Jamtaba, loop “layers” are limited to 64 beats, which is 16 measures in 4/4 time. To play a loop that is longer than 64 beats, it must be separated into parts, one per layer. All of My Loops that are longer than 64 beats have already been separated into parts for use in layers.
Assuming a 4/4 time signature, BPI (beats per interval) should be set to “number of measures times 4”. For example, for a 16 measure loop in 4/4 time, set BPI = 64. In 4/4 time, 4m = 16BPI, 8m = 32BPI, 12m = 48BPI, and 16m = 64BPI.
Start playback of the loop layer(s) at any time - it will automatically sync to the appropriate spot in the BPI countdown.
In GarageBand and Logic, import a loop MP3, then enable looping and set the looping region to match the length of the MP3. Start playback of the MP3 and ensure that output from the DAW is routed to the remote band rehearsal application.
WikiLoops is a community of musicians that record loops and share them, allowing others to use them freely for practice, to record with, etc. Most loops are full-length songs, not short loops like those usually used with remote band rehearsal applications, and are not free. If you’re looking for other musicians with which to do remote multi-track recording, this may be an option.
[END]