Image Image Image Image Image

Presented at CHI 2012, Touché is a capacitive system for pervasive, continuous sensing. Among other amazing capabilities, it can accurately sense gestures a user makes on his own body. “It is conceivable that one day mobile devices could have no screens or buttons, and rely exclusively on the body as the input surface.” Touché.

Noticing that many of the same sensors, silicon, and batteries used in smartphones are being used to create smarter artificial limbs, Fast Company draws the conclusion that the market for smartphones is driving technology development useful for bionics. While interesting enough, the article doesn’t continue to the next logical and far more interesting possibility: that phones themselves are becoming parts of our bodies. To what extent are smartphones already bionic organs, and how could we tell if they were? I’m actively researching design in this area – stay tuned for more about the body-incorporated phone.

A study provides evidence that talking into a person’s right ear can affect behavior more effectively than talking into the left.

One of the best known asymmetries in humans is the right ear dominance for listening to verbal stimuli, which is believed to reflect the brain’s left hemisphere superiority for processing verbal information.

I heavily prefer my left ear for phone calls. So much so that I have trouble understanding people on the phone when I use my right ear. Should I be concerned that my brain seems to be inverted?

Read on and it becomes clear that going beyond perceptual psychology, the scientists are terrifically shrewd:

Tommasi and Marzoli’s three studies specifically observed ear preference during social interactions in noisy night club environments. In the first study, 286 clubbers were observed while they were talking, with loud music in the background. In total, 72 percent of interactions occurred on the right side of the listener. These results are consistent with the right ear preference found in both laboratory studies and questionnaires and they demonstrate that the side bias is spontaneously displayed outside the laboratory.

In the second study, the researchers approached 160 clubbers and mumbled an inaudible, meaningless utterance and waited for the subjects to turn their head and offer either their left of their right ear. They then asked them for a cigarette. Overall, 58 percent offered their right ear for listening and 42 percent their left. Only women showed a consistent right-ear preference. In this study, there was no link between the number of cigarettes obtained and the ear receiving the request.

In the third study, the researchers intentionally addressed 176 clubbers in either their right or their left ear when asking for a cigarette. They obtained significantly more cigarettes when they spoke to the clubbers’ right ear compared with their left.

I’m picturing the scientists using their grant money to pay cover at dance clubs and try to obtain as many cigarettes as possible – carefully collecting, then smoking, their data – with the added bonus that their experiment happens to require striking up conversation with clubbers of the opposite sex who are dancing alone. One assumes that, if the test subject happened to be attractive, once the cigarette was obtained (or not) the subject was invited out onto the terrace so the scientist could explain the experiment and his interesting line of work. Well played!

Another MRI study, this time investigating how we learn parts of speech:

The test consisted of working out the meaning of a new term based on the context provided in two sentences. For example, in the phrase “The girl got a jat for Christmas” and “The best man was so nervous he forgot the jat,” the noun jat means “ring.” Similarly, with “The student is nising noodles for breakfast” and “The man nised a delicious meal for her” the hidden verb is “cook.”

“This task simulates, at an experimental level, how we acquire part of our vocabulary over the course of our lives, by discovering the meaning of new words in written contexts,” explains Rodríguez-Fornells. “This kind of vocabulary acquisition based on verbal contexts is one of the most important mechanisms for learning new words during childhood and later as adults, because we are constantly learning new terms.”

The participants had to learn 80 new nouns and 80 new verbs. By doing this, the brain imaging showed that new nouns primarily activate the left fusiform gyrus (the underside of the temporal lobe associated with visual and object processing), while the new verbs activated part of the left posterior medial temporal gyrus (associated with semantic and conceptual aspects) and the left inferior frontal gyrus (involved in processing grammar).

This last bit was unexpected, at first. I would have guessed that verbs would be learned in regions of the brain associated with motor action. But according to this study, verbs seem to be learned only as grammatical concepts. In other words, knowledge of what it means to run is quite different than knowing how to run. Which makes sense if verb meaning is accessed by representational memory rather than declarative memory.

Researchers at the University of Tampere in Finland found that,

Interfaces that vibrate soon after we click a virtual button (on the order of tens of milliseconds) and whose vibrations have short durations are preferred. This combination simulates a button with a “light touch” – one that depresses right after we touch it and offers little resistance.

Users also liked virtual buttons that vibrated after a longer delay and then for a longer subsequent duration. These buttons behaved like ones that require more force to depress.

This is very interesting. When we think of multimodal feedback needing to make cognitive sense, synchronization first comes to mind. But there are many more synesthesias in our experience that can only be uncovered through careful reflection. To make an interface feel real, we must first examine reality.

Researchers at the Army Research Office developed a vibrating belt with eight mini actuators — “tactors” — that signify all the cardinal directions. The belt is hooked up to a GPS navigation system, a digital compass and an accelerometer, so the system knows which way a soldier is headed even if he’s lying on his side or on his back.

The tactors vibrate at 250 hertz, which equates to a gentle nudge around the middle. Researchers developed a sort of tactile morse code to signify each direction, helping a soldier determine which way to go, New Scientist explains. A soldier moving in the right direction will feel the proper pattern across the front of his torso. A buzz from the front, side and back tactors means “halt,” a pulsating movement from back to front means “move out,” and so on.

A t-shirt design by Derek Eads.

Recent research reveals some fun facts about aural-tactile synesthesia:

Both hearing and touch, the scientists pointed out, rely on nerves set atwitter by vibration. A cell phone set to vibrate can be sensed by the skin of the hand, and the phone’s ring tone generates sound waves — vibrations of air — that move the eardrum…

A vibration that has a higher or lower frequency than a sound… tends to skew pitch perception up or down. Sounds can also bias whether a vibration is perceived.

The ability of skin and ears to confuse each other also extends to volume… A car radio may sound louder to a driver than his passengers because of the shaking of the steering wheel. “As you make a vibration more intense, what people hear seems louder,” says Yau. Sound, on the other hand, doesn’t seem to change how intense vibrations feel.

Max Mathews, electronic music pioneer, has died.

Though computer music is at the edge of the avant-garde today, its roots go back to 1957, when Mathews wrote the first version of “Music,” a program that allowed an IBM 704 mainframe computer to play a 17-second composition. He quickly realized, as he put it in a 1963 article in Science, “There are no theoretical limits to the performance of the computer as a source of musical sounds.”

Rest in peace, Max.

UPDATE: I haven’t updated this blog in a while, and I realized after posting this that my previous post was about the 2010 Modulations concert. Max Mathews played at Modulations too, and that was the last time I saw him.

I finally got around to recording and mastering the set I played at the CCRMA Modulations show a few months back. Though I’ve been a drum and bass fan for many years, this year’s Modulations was the first time I’d mixed it for others. Hope you like it!

Modulations 2010
Drum & Bass | 40:00 | May 2010

Audio clip: Adobe Flash Player (version 9 or above) is required to play this audio clip. Download the latest version here. You also need to have JavaScript enabled in your browser.

Download (mp3, 82.7 MB)


1. Excision — System Check
2. Randomer — Synth Geek
3. Noisia — Deception
4. Bassnectar — Teleport Massive (Bassnectar Remix)
5. Moving Fusion, Shimon, Ant Miles — Underbelly
6. Brookes Brothers — Crackdown
7. The Ian Carey Project — Get Shaky (Matrix & Futurebound’s Nip & Tuck Mix)
8. Netsky — Eyes Closed
9. Camo & Krooked — Time Is Ticking Away feat. Shaz Sparks

Over the last few days this video has been so much bombshell to many of my music-prone friends.

It’s called the Multi-Touch Light Table and it was created by East Bay-based artist/fidget-house DJ Gregory Kaufman. The video is beautifully put together, highlighting the importance of presentation when documenting new ideas.

I really like some of the interaction ideas presented in the video. Others, I’m not so sure about. But that’s all right: the significance of the MTLT is that it’s the first surface-based DJ tool that systematically accounts for the needs of an expert user.

Interestingly, even though it looks futuristic and expensive to us, interfaces like this will eventually be the most accessible artistic tools. Once multi-touch surface are ubiquitous, the easiest way to gain some capability will be to use inexpensive or open-source software. The physical interfaces created for DJing, such as Technics 1200s, are prosthetic objects (as are musical instruments), and will remain more expensive because mechanical contraptions will always be. Now, that isn’t to say that in the future our interfaces won’t evolve to become digital, networked, and multi-touch sensitive, or even that their physicality will be replaced with a digital haptic display. But one of the initial draws of the MTLT—the fact of its perfectly flat, clean interactive surface—seems exotic to us right now, and in the near future it will be default.

Check out this flexible interface called impress. Flexible displays just look so organic and, well impressive. One day these kinds of surface materials will become viable displays and they’ll mark a milestone in touch computing.

It’s natural to stop dancing between songs. The beat changes, the sub-rhythms reorient themselves, a new hook is presented and a new statement is made. But stopping dancing between songs is undesirable. We wish to lose ourselves in as many consecutive moments as possible. The art of mixing music is to fulfill our desire to dance along to continuous excellent music, uninterrupted for many minutes (or, in the best case, many hours) at a time. (Even if we don’t explicitly move our bodies to the music, when we listen our minds are dancing; the same rules apply.)

I don’t remember what prompted me to take that note, but it was probably not that the mixing was especially smooth.



A tomato hailing from Capay, California.

LHCSound is a site where you can listen to sonified data from the Large Hadron Collider. Some thoughts:

  • That’s one untidy heap of a website. Is this how it feels to be inside the mind of a brilliant physicist?
  • The name “LHCSound” refers to “Csound”, a programming language for audio synthesis and music composition. But how many of their readers will make the connection?
  • If they are expecting their readers to know what Csound is, then their explanation of the process they used for sonification falls way short. I want to know the details of how they mapped their data to synthesis parameters.
  • What great sampling material this will make. I wonder how long before we hear electronic music incorporating these sounds.

The Immersive Pinball demo I created for Fortune’s Brainstorm:Tech conference was featured in a BBC special on haptics.

I keep watching the HTC Sense unveiling video from Mobile World Congress 2010. The content is pretty cool, but I’m more fascinated by the presentation itself. Chief marketing officer John Wang gives a simply electrifying performance. It almost feels like an Apple keynote.

The iFeel_IM haptic interface has been making rounds on the internet lately. I tried it at CHI 2010 a few weeks ago and liked it a lot. Affective (emotional haptic) interfaces are full of potential. IFeel_IM mashes together three separate innovations:

  • Touch feedback in several different places on the body: spine, tummy, waist.
  • Touch effects that are generated from emotional language.
  • Synchronization to visuals from Second Life

All are very interesting. The spine haptics seemed a stretch to me, but the butterfly-in-the-tummy was surprisingly effective. The hug was good, but a bit sterile. Hug interfaces need nuance to bring them to the next level of realism.

The fact that the feedback is generated from the emotional language of another person seemed to be one of the major challenges—the software is built to extract emotionally-charged sentences using linguistic models. For example, if someone writes “I love you” to you, your the haptic device on your tummy will react by creating a butterflies-like sensation. As an enaction devotee I would rather actuate a hug with a hug sensor. Something about the translation of words to haptics is difficult for me to accept. But it could certainly be a lot of fun in some scenarios!

I’ve re-recorded my techno mix Awake with significantly higher sound quality. So if you downloaded a copy be sure to replace it with the new file!

Awake

Awake
Techno | 46:01 | October 2009

Audio clip: Adobe Flash Player (version 9 or above) is required to play this audio clip. Download the latest version here. You also need to have JavaScript enabled in your browser.

Download (mp3, 92 MB)


1. District One (a.k.a. Bart Skils & Anton Pieete) — Dubcrystal
2. Saeed Younan — Kumbalha (Sergio Fernandez Remix)
3. Pete Grove — I Don’t Buy It
4. DBN — Asteroidz featuring Madita (D-Unity Remix)
5. Wehbba & Ryo Peres — El Masnou
6. Broombeck — The Clapper
7. Luca & Paul — Dinamicro (Karotte by Gregor Tresher Remix)
8. Martin Worner — Full Tilt
9. Joris Voorn — The Deep

I recently started using Eclipse on OS X and it was so unresponsive, it was almost unusable. Switching tabs was slow, switching perspectives was hella slow. I searched around the web for a solid hour for how to make it faster and finally found the solution. Maybe someone can use it.

My machine is running OS X 10.5, and I have 2 GB of RAM. (This is important because the solution requires messing with how Eclipse handles memory. If you have a different amount of RAM, these numbers may not work and you’ll need to fiddle with them.)

  • Save your work and quit Eclipse.
  • Open the Eclipse application package by right-clicking (or Control-clicking) on Eclipse.app and select “Show Package Contents.”
  • Navigate to Contents→MacOS→, and open “eclipse.ini” in your favorite text editor.
  • Edit the line that starts with -”XX:MaxPermSize” to say “-XX:MaxPermSize=128m”.
  • Before that line, add a line that says “-XX:PermSize=64m”.
  • Edit the line that starts with “-Xms” to say “-Xms40m”.
  • Edit the line that starts ith “-Xmx” to say “-Xmx768m”.
  • Save & relaunch Eclipse.

Worked like a charm for me.

Scroll to Top

To Top

musical instruments

11

May
2011

No Comments

In history
interfaces
music

By Dave

“He imagined and created his own magical world and first built the essential concepts and tools that allowed us all to do the same.”

On 11, May 2011 | No Comments | In history, interfaces, music | By Dave

Max Mathews, electronic music pioneer, has died.

Though computer music is at the edge of the avant-garde today, its roots go back to 1957, when Mathews wrote the first version of “Music,” a program that allowed an IBM 704 mainframe computer to play a 17-second composition. He quickly realized, as he put it in a 1963 article in Science, “There are no theoretical limits to the performance of the computer as a source of musical sounds.”

Rest in peace, Max.

UPDATE: I haven’t updated this blog in a while, and I realized after posting this that my previous post was about the 2010 Modulations concert. Max Mathews played at Modulations too, and that was the last time I saw him.

Tags | , , ,

20

Aug
2010

No Comments

In interfaces
music

By Dave

Luscious surface for DJ performance

On 20, Aug 2010 | No Comments | In interfaces, music | By Dave

Over the last few days this video has been so much bombshell to many of my music-prone friends.

It’s called the Multi-Touch Light Table and it was created by East Bay-based artist/fidget-house DJ Gregory Kaufman. The video is beautifully put together, highlighting the importance of presentation when documenting new ideas.

I really like some of the interaction ideas presented in the video. Others, I’m not so sure about. But that’s all right: the significance of the MTLT is that it’s the first surface-based DJ tool that systematically accounts for the needs of an expert user.

Interestingly, even though it looks futuristic and expensive to us, interfaces like this will eventually be the most accessible artistic tools. Once multi-touch surface are ubiquitous, the easiest way to gain some capability will be to use inexpensive or open-source software. The physical interfaces created for DJing, such as Technics 1200s, are prosthetic objects (as are musical instruments), and will remain more expensive because mechanical contraptions will always be. Now, that isn’t to say that in the future our interfaces won’t evolve to become digital, networked, and multi-touch sensitive, or even that their physicality will be replaced with a digital haptic display. But one of the initial draws of the MTLT—the fact of its perfectly flat, clean interactive surface—seems exotic to us right now, and in the near future it will be default.

Tags | , , ,

09

Oct
2009

No Comments

In art
music
perception

By David Birnbaum

Incredible speaking piano

On 09, Oct 2009 | No Comments | In art, music, perception | By David Birnbaum

A composer named Peter Ablinger has created a jaw dropping sound art piece. He recorded a speech read by a child, analyzed the recording to extract its frequency content, and then mapped it to pitches on an acoustic player piano. My reaction was identical to the one described in the interview: what at first sounds like nonsense comes into perfect focus when you begin reading the text along to the sound. The flip from unintelligibility to clarity is a thrilling experience. Beautiful, beautiful work!

Tags | ,

The Hand

On 04, Jun 2009 | One Comment | In art, books, cognition, neuroscience, physiology, tactility | By David Birnbaum

0679740473The Hand by Frank Wilson is a rare treat. It runs the gamut from anthropology (both the cultural and evolutionary varieties), to psychology, to biography. Wilson interviews an auto mechanic, a pupeteer, a surgeon, a physical therapist, a rock climber, a magician, and others—all with the goal of understanding the extent to which the human hand defines humanness.

Wilson is a neurologist who works with musicians who have been afflicted with debilitating chronic hand pain. As he writes about his many interviews, a few themes emerge that are especially relevant to our interests here.

Incorporation
Incorporation is the phenomenon of internalizing external objects; it’s the feeling that we all get that a tool has become one with our body.

The idea of “becoming one” with a backhoe is no more exotic than the idea of a rider becoming one with a horse or a carpenter becoming one with a hammer, and this phenomenon itself may take its origin from countless monkeys who spent countless eons becoming one with tree branches. The mystical feel comes from the combination of a good mechanical marriage and something in the nervous system that can make an object external to the body feel as if it had sprouted from the hand, foot, or (rarely) some other place on the body where your skin makes contact with it…

The contexts in which this bonding occurs are so varied that there is no single word that adequately conveys either the process or the many variants of its final form. One term that might qualify is “incorporation”—bringing something into, or making it part of, the body. It is a commonplace experience, familiar to anyone who has ever played a musical instrument, eaten with a fork or chopsticks, ridden a bicycle, or driven a car. (p. 63)

Projection
Projection is the ability to use the hand as a bridge for projecting consciousness from one location to another. (Wilson did not use the word “projection” in the book.) In some ways, projection can be seen as the opposite of incorporation. Master puppeteer Anton Bachleitner:

It takes at least three years of work to say you are a puppeteer. The most difficult job technically is to be able to feel the foot contact the floor as it actually happens. The only way to make the puppet look as though it is actually walking is by feeling what is happening through your hands. The other thing which I think you cannot really train for, but only can discover with very long practice and experience, is a change in your own vision.

The best puppeteer after some years will actually see what is happening on the stage as if he himself was located in the head of the puppet, looking out through the puppet’s eyes—he must learn to be in the puppet. This is true not only in the traditional actor’s sense, but in an unusual perceptual sense. The puppeteer stands two meters above the puppet and must be able to see what is on the stage and to move from the puppet’s perspective. Moving is a special problem because of this distance, because the puppet does not move at the same time your hand does. Also, there can be several puppets on the stage at the same time, and to appear realistic they must react to each other as they would in real life. So again the puppeteer must himself be mentally on the stage and able to react as a stage actor would react. This is something I cannot explain, but it is very imprortant for a puppeteer to be able to do this. (pp. 92–93)

Serge Percelly, professional juggler:

[An act is successful] not because you put something in the act that’s really difficult, but because you put something in the act in exactly the right way—in a way that makes it more interesting, not only for me but for the audience as well. I’m just trying somehow to do the act that I would have loved to see. (p. 111)

Skill
Wilson is a musician and a doctor to musicians, so he has special insight into the neurology of musical skill—which he recognizes as special case of manual skill that involves gesture, communication, and emotion.

Musical skill provides the clearest example and the cleanest proof of the existence of a whole class of self-defined, personally distinctive motor skills with an extended training and experience base, strong ties to the individual’s emotional and cognitive development, strong communicative intent, and very high performance standards. Musical skill, in other words, is more than simply praxis, ordinary manual dexterity, or expertness in pantomime. (p. 207)

The upper-limb (or “output”) requirements for an instrumentalist are not unique either; they depend upon the possession of arms, fingers, and thumbs, specific but idiosyncratic limits on the rage of motion at the shoulder, elbow, wrist, hand, and finger joints, variable abilities to achieve repetition rates and forces with specific digital configurations in sequence at multiple contact points on a sound-making device, and so on. Peculiarities in the physical configuration and movement capabisities of the musician’s limbs can be an advantage or disadvantage but are reflected in (and in adverse cases can be overcome through) instrument design: How wide can you make the neck of a guitar? How far apart should the keys be on a piano? Where should the keys be placed on a flute—in general? and for Susan and Peter? (p. 225)

Awareness
Touch experience can be a gateway to awareness, which can in turn heal both the mind and the body. Moshe Feldenkrais invented a form of physical therapy that focuses on stimulating an awareness of touch and movement sensations in order to relieve pain.

Most people slouch, tilt, shuffle, twist, stumble, and hobble along. Why should that be? Was there something wrong with their brains? After considering what dancers and musicians go through to improve control of their movements, [Feldenkrais] guessed that people must either be ignorant of the possibilities or refuse to act on them. So they just heave themselves around, lurching from parking place to office to parking place, utterly oblivious to what they are doing, to their appearance, and even to the sensations that arise from bodily movement. He suspected that people just lose contact with their own bodies. If and when they do notice, it is because they are so stiff that they can’t get out of bed or are in so much pain that they can barely get out of a chair. Then they start noticing…

What [Feldenkrais] was doing did not seem complicated. The goal of the guided movements was not to learn how to move, in the sense of learning to do a new dance step. The goal was not to stretch ligaments or muscles. It was not to increase strength. The goal, as he saw it, was to get the messages moving again and to encourage the brain to pay attention to them. (p. 244)

And his student, Anat Baniel, on the deep psychological roots of movement disorders:

I think working with children has given me this idea, which isn’t often discussed in medicine: a lot of disease—medical disease and emotional “dis-ease”—is an outcome of a lack of full development. It’s not something we can get to just by removing a psychological block…

Of course there are problems due to traumatic events in childhood, or disease—you name it. Feldenkrais said that ideal development would happen if the child was not opposed by a force too big for its strength. When you say to a small child, “Don’t touch that, it’s dangerous!” you create such a forceful inhibition that you actually distort the child’s movement, and growth, in a certain way.

Feldenkrais taught us to look for what isn’t there. Why doesn’t movement happen in the way that it should, given gravity, given the structure of the body, given the brain? For all of us there is a sort of sphere, or range, of movement that should be possible. Some people get only five or ten percent of that sphere, and you have to ask, “What explains the difference between those who get very little and those who get a lot?” Feldenkrais said that the difference is that in the process of development, the body encountered forces that were disproportionate to what the nervous system could absorb without becoming overinhibited—or overly excited, which is a manifestation of the same thing. (p. 252)

Feldenkrais’s approach is fascinating, but there is scant discussion in Wilson’s book about the role of the therapist’s hand in this process. After all, this kind of therapy is wholly reliant on an accidental discovery: that the patient can be made aware of her own body through an external, expert hand radiating pressure and heat. How is this possible? The topic isn’t explored.

There are many, many wonderful things to learn from this book for anyone with an interest in biology, art, music, history, or sports. You can find Frank Wilson on the web at Handoc.com

Tags | , , ,

01

Sep
2008

No Comments

In interfaces
music

By David Birnbaum

Automatically sync your music to your body

On 01, Sep 2008 | No Comments | In interfaces, music | By David Birnbaum

That’s what Yamaha’s BODiBEAT promises to do. More specifically, it syncs the tempo of your music to your gait. It’s been released as a workout tool, but it’s also an interesting musical interface as such, and if it works it will no doubt find its way into electronic music performance very soon.

(via Engadget)

Tags | ,

15

Jun
2008

No Comments

In music

By David Birnbaum

Moog smacks down Gibson in guitar-improvement-off

On 15, Jun 2008 | No Comments | In music | By David Birnbaum

I blogged Gibson’s pretentiously named “Robot Guitar” a while back. I haven’t used it yet, but this video makes me think its interaction design is questionable.

Donald Norman might break it down like this:

  • Pull out the “tuning knob,” which looks like a standard gain control and doesn’t look like it’s supposed to be pulled. In fact, when the knobs on normal guitars are pulled in this way they come off, so most guitarists have learned to avoid pulling on knobs.
  • Make sure you’re not touching the neck or the tuning keys! You know, those things that look like they were made for human fingers to touch and that are an essential part of the manual interface to all other string instruments? Forget that, and instead keep this knowledge in your head: hands off.
  • Strum the guitar several times. While you do this, crouch over your guitar so you can see the LEDs on the front of the tuning knob, which faces the audience.
  • After a few strums, the LEDs change color. Now you’re “ready to rock.”

Meanwhile, Moog has extended the creative capabilities of electric guitars with innovative, elegant technology that may actually deserve the title “world’s first,” though Moog is confident enough in their product that they don’t bother with that kind of marketing nonsense. By feeding some of the output signal back into the strings, the instrument allows precise control of the envelope of the sound. “Infinite sustain” as well as banjo-like damping are both made possible. Powerful stuff.

Tags | , , ,

18

Apr
2008

No Comments

In books

By David Birnbaum

Singulatarianism

On 18, Apr 2008 | No Comments | In books | By David Birnbaum

TSiN

Ray Kurzweil’s The Singularity Is Near is worth reading. It’s a great introduction to the “emerging technologies” meme. If you’ve already begun wandering down the diabolical path of contemplating the obsolescence of biology and the societal transformation that will result, Singularity doesn’t reveal much. However it’s an easy read and a cultural milestone, so pushing through it is still probably a good idea.

That said, I had a few gripes with the book. Firstly, the author’s tone doesn’t sound like the cool scientific theorist he claims to be. Virtually every discussion of progress in biotechnology lends particular weight to the synthesizing of artificial pancreas cells, which Kurzweil readily acknowledges he needs to reverse the course of his diabetes. He repeats over and over that people (like him) who take full advantage of today’s knowledge about longevity will live to see the Singularity (and by corollary, live indefinitely), and that those who do not will needlessly pay the price of annihilation. I don’t remember even one acknowledgment that his own efforts may fail (or be misguided!) despite his dogged adherence to health guidelines derived from cutting edge research. This doesn’t seem to reflect the attitude of a detached intellectual.

Maybe I’m being too harsh. It’s pretty clear that Kurzweil tries to make his subject singularly exciting, and the book seems to be written at least partially for the journalists who might like to capitalize on the author’s theory by writing articles to the effect of, “This man says you’ll socialize with robots one day. Isn’t that insane plus totally interesting?”

Still, I found his discussions of sociology and morality hugely lacking. Kurzweil falls far short in his brief examination of the coming collisions between the Singularity and virtually every modern social structure. He might counter that his goal for the book was to cultivate optimism and open minds to the grand potential of the coming transcendence. But with its foreboding title, one would expect the book to not only open minds but also put them on guard against perversions of his grand vision. And for his slight allowance that catastrophe may wipe us all out before (or after) the Singularity, he does not explicitly acknowledge that our place in the universe is so absurdly minute that, for instance, even after our entire solar system becomes a giant computer the collision of two unknown (or even known) neutron stars halfway across our galaxy could fry it all, and human technology might be powerless to stop it.

I extracted some passages that were relevant to haptics:

The current disadvantages of Web-based commerce (for example, limitations in the ability to directly interact with products and the frequent frustrations of interacting with inflexible menus and forms instead of human personnel) will gradually dissolve as the trends move robustly in favor of the electronic world. By the end of this decade, computers will disappear as distinct physical objects, with displays built in our eyeglasses and electronics woven in our clothing, providing full-immersion visual virtual reality. Thus, “going to a Web site” will mean entering a virtual-reality environment—at least for the visual and auditory senses—where we can directly interact with products and people, both real and simulated… Haptic (tactile) interfaces will enable us to touch products and people. It is difficult to identify any lasting advantage of the old brick-and-mortar world that will not ultimately be overcome by the rich interactive interfaces that are soon to come. (104–105)

I don’t like “haptic (tactile)” (they’re not equivalent), but the point is clear. There were also some references to musical interaction research:

Edward Taub at the University of Alabama studied the region of the cortex responsible for evaluating the tactile input from the fingers. Comparing non-musicians to experienced players of stringed instruments, he found no difference in the brain regions devoted to the fingers of the right hand but a huge difference for the fingers of the left hand. (174)

Dr. Alvaro Pascual-Leone at Harvard University scanned the brains of volunteers before and after they practiced a simple piano exercise. The brain motor cortex of the volunteers changed as a direct result of their practice. He then had a second group just think about doing the piano exercise but without actually moving any muscles. This produced an equally pronounced change in the motor-cortex network. (175)

Finally, I think the use of the word “mastering” here brings up an important point:

Machines have exacting memories. Contemporary computers can master billions of facts accurately, a capability that its doubling every year. (261)

Haptic metaphors like this are used generously throughout the book, no doubt to underscore Kurzweil’s intense belief that humanity is ingrained in nonbiological intelligent technology. However, as a terminology freak I would argue that computers memorize facts, but don’t master them. Mastery comes with an ability to manipulate freely. For mastery, you need a body.

Tags | , , , ,

27

Feb
2008

One Comment

In music

By David Birnbaum

Digital Orchestra concert approaches

On 27, Feb 2008 | One Comment | In music | By David Birnbaum

Rulers in DCS

The McGill University Digital Orchestra is putting on their first performance on March 5th at 7:30 p.m. in Montreal’s Pollack Hall. The lovely and talented Xenia Pestova will be playing the Rulers, an instrument I invented. The piece she will be playing, which I haven’t heard yet, was composed by D. Andrew Stewart and is the first music to be written for the instrument. There will be a live webcast of the show. To watch it, launch the QuickTime application (free download) a few minutes before the start of the show, select “Open URL in New Player” from the File menu, and enter: rtsp://132.206.142.8/pollackhall.sdp

Tags | , ,

11

Dec
2007

One Comment

In music
robotics

By David Birnbaum

Problems and prospects for Gibson’s self-tuning guitar

On 11, Dec 2007 | One Comment | In music, robotics | By David Birnbaum

Gibson has announced a guitar with a built-in self-tuning mechanism. Some have suggested that there is a problem with allowing people to skip learning how to tune a guitar before they play it, because tuning helps develop the ear. I think this is a valid concern, and readers of my papers would know I don’t think lowering the entry fee for musical instrumental interaction is, in itself, a “good thing.” At the same time, there are plenty of advantages offered by a self-tuning guitar that have nothing at all to do with ear training, such as avoiding the need to bring a capo to gigs, or to bring more than one guitar to a show for quickly playing two consecutive songs than require drastically different guitar tunings. (Besides, there are plenty of other excellent ways to train your ear.) Quick but accurate tuning changes will also surely be exploited in composition; tuning changes can be done in the middle of a piece, and the musical capabilities and quirkiness of the auto-tuner could even be used for some as-yet-unknown artistic end.

What I find especially interesting is how the words “world’s first robot guitar” are tossed around in this press release. First of all, it seems as if the word “robot” is being used vaguely to refer to the presence of a servo system. If this guitar is robotic, then so is my laptop for its ability to read and eject optical media. I think we’re going to see more of this, similar to the way “net” was overused in the nineties. We are entering a robo-sheik era where any product that can possibly justify doing so will be incorporating the word “robot” into its name.

As for the “world’s first” claim, someone should tell Gibson about TransPerformance, the company that has already been selling automatic tuner retrofits since 2005, as well as the dozens of other music technology projects that are based on guitar interaction and involve motors. It’s old, but anyone who hasn’t yet seen the League of Musical Urban Robots (LEMUR) video of the LEMUR Guitar Bot should check it out:

It’s easier for me to accept calling the LEMUR Guitar Bot a “robot” than the Gibson self-tuner. What do you think?

Tags | , ,

14

Sep
2007

No Comments

In outreach

By David Birnbaum

The Rulers @ NextFest

On 14, Sep 2007 | No Comments | In outreach | By David Birnbaum

Rulers (played)

I’m excited to announce that, as I write this post, a musical instrument that I designed and built is being shown at WIRED magazine’s NextFest conference and trade show in Los Angeles. The first version of the Rulers came out of a physical interaction design workshop hosted by Stanford University’s CCRMA music technology program. Later I gave it a full version upgrade and it was purchased by the Digital Orchestra at McGill University’s Schulich School of Music. Since the Digital Orchestra is at NextFest, so are the Rulers. Here’s the press release for the exhibition.

Tags | , , ,

21

Jun
2007

No Comments

In music

By David Birnbaum

Rulers: First contact with composer and performer

On 21, Jun 2007 | No Comments | In music | By David Birnbaum

Last week the Rulers v1.0 were unveiled to composer D. Andrew Stewart who will be writing Rulers music, and musician Xenia Pestova who will be performing the instrument next spring. The meeting went very well—the artists found the interface visually and tactually inspiring, which would have been obvious by the way they played with it even if they had not told me so. I took some video of the encounter, posted below. Note that this was the very first time that the interface was hooked up to sound software, so the mapping and physical modeling synth were thrown together to get something working. In other words, the sound in the video doesn’t represent how the system will behave in the end. However I think it’s still pretty clear that the interface is highly responsive to instrumental gestures. I’ll be posting more as software, musical exercises, and eventually pieces are written for it.

Update: I’m having issues with the video… apparently Sony’s MPG isn’t being recognized by any video playback application I’ve tried yet except for the Finder preview. YouTube and Motionbox both don’t recognize them either. I hope I will solve the problem soon.

Tags | ,

21

Jun
2007

No Comments

In interfaces
music

By David Birnbaum

Why does Arduino have to be so fabulous?

On 21, Jun 2007 | No Comments | In interfaces, music | By David Birnbaum

Converting analog voltages to computer data has been a central part of my life for four years now. Back in the dark ages, I used the Atomic Pro, a $1500 plastic white box. In 2005, my colleague Mark Marshall developed the AVR-HID. I have built upwards of five of them, and they work really well. But a few months back I tried using Arduino, and it quickly stole my heart: eight channels of 10-bit A/D on a PCB the size of a twoonie. More importantly, it comes with a free lightweight application for programming the microcontroller, which I prefer to using the Terminal. I moved the Rulers to the Arduino platform and will use it again for my Breakflute.

Tags | , , ,

22

Mar
2007

No Comments

In music

By David Birnbaum

Hang

On 22, Mar 2007 | No Comments | In music | By David Birnbaum

My Powerplant Family partner Lucy May just told me she is going to be choreographing a piece to be played on a Hang drum, a pitched metal instrument from Berne, Switzerland invented in 2000. Apparently these things are all the rage in the Netherlands right now. Here’s a video of a Hang player jamming… not in a laboratory or high tech concert hall, but at a picnic. Who ever heard of music being played at a picnic?! Impossible.

Tags | ,