France Atlanta 2020 : Virtual Exhibitions and webinars - Machines of Loving Grace
All Watched Over by Machines of Loving Grace is a 1967 poem by Richard Brautigan that describes the peaceful and harmonious cohabitation of humans and computers. Fifty years later – following the technological revolutions brought about by the internet, big data, cloud computing and deep learning – Brautigan’s vision resonates singularly. It compels us to rethink our interactions with machines, even in regards to the intimate act of artistic creation. In collaboration with its American partners, France’s Institute for Research and Coordination in Acoustics/Music (IRCAM) invites participants to discover some remarkable examples of possible artistic relationships between humans and computers during two seminars featuring researchers and artists who are building this new deal.
This event is organized by IRCAM – the Institute for Research and Coordination in Acoustics/Music – and the STMS lab (IRCAM, The French National Centre for Scientific Research, the Ministry of Culture and Sorbonne University), in collaboration with Georgia Tech’s School of Music, NYU Steinhardt – Music Education, the UC Berkeley Center for New Music and Audio Technologies, and the Atlanta Office of the Cultural Services of the Embassy of France in the United States.
As part of the 2020 France Atlanta Festival.
- "Working Creatively with Machines", October 14, 11:00am (Atlanta) / 5:00 pm (Paris) : Open discussion on composition and writing.
Introduction by Yves Berthelot, Vice-Provost for International Initiatives, Georgia Institute of Technology and Frank Madlener, director of IRCAM.
Free / Review streaming
- "Performances with Machines", October 15, 11am (Atlanta) / 5:00pm (Paris) : Open discussion on performance.
Free / Review streaming
From October 12 to November 10, a virtual exhibition is accessible in order to present excerpts of works and tools that heavily rely on artificial intelligence techniques and that question the relationship between human and machines. The exhibition includes works by such artists and researchers as Daniele Ghisi (composer), Jérôme Nika and Gérard Assayag (researchers, IRCAM), Jason Freeman (Georgia Tech), Carmine Emanuele Cella (professor and composer, CNMAT – CU Berkeley), Alex Ruthman (professor, New York University).
- Daniele Ghisi: Born in Italy in 1984, Daniele Ghisi studied Music Composition at Bergame Conservatory with S. Gervasoni and continued his studies with IRCAM Cursus. In 2009-2010, he is a residency composer at Akademie der Künste (Berlin), in 2011-2012, he is residency composer in Spain, member of the Academie de France in Madrid – Casa de Velázquez. In 2015, he is in residency in Milano, with the Divertimento Ensemble, which recorded it’s first monographic CD (Geografie). Since 2010, he develops, with composer Andrea Agostini, “bach: automated composer’s helper”, the library for computer-assisted composition. He is the co-founder of the blog nothing.eu, in which he writes. He is edited by Ricordi. Between 2017 and 2020, he teaches Electroacoustic Composition at the Gênes Conservatory. Actually, he’s composer-researcher at California University, Berkeley (CNMAT).
- Elaine Chew: Elaine Chew is a senior CNRS researcher at the STMS Lab at IRCAM and a Visiting Professor at King's College London. Her research focuses on mathematical and computational modeling of musical structures in music and in electrocardiographic sequences. The tools are used to explicate leading performances, understand music-heart-brain interactions, stratify arrhythmias into subtypes, and generate AI music. Her work has been recognised by PECASE and NSF CAREER awards, Fellowships at Harvard's Radcliffe Institute for Advanced Study, and ERC ADG and POC awards. She was Professor of Digital Media at QMUL (2011-2019), Assistant/tenured Associate Professor and Viterbi Early Career Chair at USC (2001-2013). and held visiting appointments at Harvard (2008-2009) and Lehigh (2000-2001). As a pianist, she integrates her research into concert-conversations that showcase scientific visualisations and lab-grown compositions. She received PhD and SM degrees in Operations Research from MIT, BAS in Mathematical & Computational Sciences (honors) and Music (distinction) from Stanford, and LTCL/FTCL piano diplomas.
- Jérome Nika: Jérôme Nika is researcher in human-machine musical interaction in the Music Representations Team at Ircam. He graduated from the French Grandes Écoles Télécom ParisTech and ENSTA ParisTech. In addition, he studied acoustics, signal processing and computer science applied to music, and composition. He specialized in the applications of computer science and signal processing to digital creation and music through a PhD (Young Researcher Prize in Science and Music, 2015; Young Researcher Prize awarded by the French Association of Computer Music, 2016), and then as a researcher at Ircam. His research focuses on the introduction of authoring, composition, and control in human-computer music co-improvisation. This work led to numerous collaborations and musical productions, particularly in improvised music (Steve Lehman, Bernard Lubat, Benoît Delbecq, Rémi Fox) and contemporary music (Pascal Dusapin, Marta Gentilucci). In 2019 – 2020, his work was involved in 3 ambitions productions : Lullaby Experience, an evolutive project by composer Pascal Dusapin, and two improvised music projects: Silver Lake Studies, in duo with Steve, and C’est Pour ça, in duo with Rémi Fox. In 2020 he is in residency at Le Fresnoy – Studio National des Arts Contemporains. More info: https://jeromenika.com
- Gérard Assayag: Gerard Assayag, an Ircam research director, is head of IRCAM Music Representation Team, a team he has founded in 1992. He has been head of the Ircam research lab (STMS, a joint lab between Ircam, CNRS and Sorbonne University) between 2011 and 2017. His research interests are centered on music representation issues, including programming languages, machine learning, constraint and visual programming, computational musicology, music modeling, and computer-assisted composition and interaction. He has designed with his collaborators OpenMusic and OMax, two music research software environments which have gained international reputation and are used in many places for computer assisted composition, analysis and improvisation.
- Jason Freeman: Jason Freeman is a Professor of Music at Georgia Tech and Chair of the School of Music. His artistic practice and scholarly research focus on using technology to engage diverse audiences in collaborative, experimental, and accessible musical experiences. He also develops educational interventions in K-12, university, and MOOC environments that broaden and increase engagement in STEM disciplines through authentic integrations of music and computing. His music has been performed at Carnegie Hall, exhibited at ACM SIGGRAPH, published by Universal Edition, broadcast on public radio’s Performance Today, and commissioned through support from the National Endowment for the Arts. Freeman’s wide-ranging work has attracted over $10 million in funding from sources such as the National Science Foundation, Amazon, and Turbulence. It has been disseminated through over 80 refereed book chapters, journal articles, and conference publications. Freeman received his B.A. in music from Yale University and his M.A. and D.M.A. in composition from Columbia University.
- Carmine Emanuele Cella: Carmine Emanuele Cella is a internationally renown composer with advanced studies in applied mathematics. He studied at Conservatory of Music G. Rossini in Italy getting master degrees in piano, computer music and composition and he got a PhD in musical composition at the Accademia di S. Cecilia in Rome. He also studied philosophy and mathematics and got a PhD in mathematical logic at the University of Bologna entitled On Symbolic Representations of Music (2011). In 2007-2008, Carmine-Emanuele Cella works as a researcher in Paris in Ircam’s Analysis/Synthesis team working on audio indexing and since January 2019, he is Assistant Professor of Music and Technology at the University of Berkeley.
- Alex Ruthmann: S. Alex Ruthmann is Associate Professor of Music Education & Music Technology, and the Director of the NYU Music Experience Design Lab (MusEDLab) at NYU Steinhardt. He holds affiliate appointments with the NYU Digital Media Design for Learning program, and the Program on Creativity and Innovation at NYU Shanghai. He currently serves as Chair of the Music in Schools and Teacher Education Commission for the International Society for Music Education. Ruthmann is co-author of Scratch Music Projects, a new book published by Oxford University Press bringing creative music and coding projects to students and educators. He is co-editor of the Oxford Handbook of Technology and Music Education, and the Routledge Companion to Music, Technology and Education. He also serves as Associate Editor of the Journal of Music, Technology, and Education. Ruthmann’s research focuses on the design of new technologies and experiences for music making, learning and engagement. Partners include the New York Philharmonic, Shanghai Symphony, Peter Gabriel, Herbie Hancock, Yungu and Portfolio Schools, Tinkamo, UNESCO, and the Rock and Roll Forever Foundation. The MusEDLab creative learning and software projects are in active use by over 900,000 people in over 150 countries across the world.
- Rémy Mignot: Rémi Mignot is a researcher in the analysis-synthesis team of IRCAM. In 2009, he obtained his PhD from the Sorbonne University, on the modeling and simulation of acoustic waves for wind instruments, with Thomas Hélie (IRCAM) and Denis Matignon (Supaero). In 2010-2012, he joined the Langevin Institut (Paris) for post-doctoral research on the sampling of room impulse responses using compressed sensing, with Laurent Daudet (Paris Diderot) and François Ollivier (UPMC). From 2012-2014, he worked with Vesa Välimäki on extended subtractive synthesis for musical instruments in the department of Signal Processing and Acoustics of Aalto University at Espoo (Finland). He returned to IRCAM in 2014 to research audio indexing and classification with Geoffroy Peeters. Since 2018, he has been responsible for research on music information retrieval in the analysis-synthesis team.
- Nicolas Obin: Nicolas Obin is a researcher in audio signal processing, machine learning, and statistical modeling of sound signals with specialization on speech processing. My main area of research is the generative modeling of the expressivity in spoken and singing voices, with application to various fields such as speech synthesis, conversational agents, and computational musicology. He is actively involved in promoting digital science and technology for arts, culture, and heritage. In particular, he collaborated with renowned artists (Georges Aperghis, Philippe Manoury, Roman Polansky, Philippe Parreno, Eric Rohmer, André Dussolier), and helped to reconstitute the digital voice of personalities, like the artificial cloning of André Dussolier’s voice (2011), the short-film Marilyn (P. Parreno, 2012) and Juger Pétain documentary (R. Saada, 2014).He regularly conducts guest lectures for reknown schools (Collège de France, Ecole Normale Supérieure, Sciences Po), organizations (CNIL, AIPPI) and in the press and the media (Le Monde, Télérama, TF1, France 5, Arte, Pour la Science).
- Grace Leslie: Grace Leslie is a flutist, electronic musician, and scientist. She develops brain-music interfaces and other physiological sensor systems that reveal aspects of her internal cognitive and affective state, those left unexpressed by sound or gesture, to an audience. As an electronic music composer and improviser, she maintains a brain-body performance practice. Grace strives to integrate the manners of conventional emotional and musical expression that she learned as a flutist with the new forms available to her as an electronic musician, using brain- computer interface to reveal aspects of her internal mental state, those left unexpressed by sound or gesture, to an audience. In recent years she has performed this music in academic and popular music venues, conferences and residencies in the United States, UK, Australia, Germany, Singapore, South Korea, China, and Japan, and released three records of this mind-body music. During her Ph.D. (Music and Cognitive Science at UCSD) studies she completed a yearlong position at Ircam in Paris, where she collaborated on an interactive sound installation and performed experiments studying the effect of active involvement on music listening. She completed her undergraduate and Masters work in Music, Science, and Technology at CCRMA,
- Benjamin Levy: Computer music designer at IRCAM, Benjamin Lévy studied both sciences—primarily computer science, with at PhD in engineering—and music. Since 2008, he has collaborated on both scientific and musical projects with several teams at IRCAM, in particular around the OMax improvisation software. As an R&D engineer and developer, he has also worked in the private sector for companies specialized in creative audio technologies. He has taken part in several artistic projects at IRCAM and elsewhere as a computer musician for contemporary music works as well as jazz, free improv, theater, and dance. He has collaborated with choreographers such as Aurélien Richard, worked on musical theater with Benjamin Lazar, and performs with the jazz saxophonist Raphaël Imbert.
- Philippe Esling: Philippe Esling received an MSc in Acoustics, Signal Processing and Computer Science in 2009 and a PhD on multiobjective time series matching in 2012. He was a post-doctoral fellow in the department of Genetics and Evolution at the University of Geneva in 2012. He is now an associate professor with tenure at IRCAM, Paris 6 since 2013. In this short time span, he authored and co-authored over 15 peer-reviewed journal papers in prestigious journals such as ACM Computing Surveys, Publications of the National Academy of Science, IEEE TSALP and Nucleic Acids Research. He received a young researcher award for his work in audio querying in 2011 and a PhD award for his work in multiobjective time series data mining in 2013. In applied research, he developed and released the first computer-aided orchestration software called Orchids, commercialized at fall 2014 and already used by a wide community of composer. He directed six Masters interns, a C++ developer for a full year, and is currently directing two PhD students. He is the lead investigator in time series mining at IRCAM, main collaborator in the international France-Canada SSHRC partnership and the supervisor of an international workgroup on orchestration.
- IRCAM: The Institute for Research and Coordination in Acoustics/Music is one of the world’s largest public research centers dedicated to musical creation and scientific research. A unique venue where artistic vision converges with scientific and technological innovation, the institute directed by Frank Madlener brings together over 160 collaborators. IRCAM hosts the UMR9912 STMS Ircam – CNRS – Sorbonne University science and technologies research lab.
- STMS Lab: The fundamental principle of the STMS Lab is to encourage productive interaction among scientific research, technological developments, and contemporary music production. Since its establishment in 1995, this initiative has provided the foundation for the institute’s activities. One of the major issues is the importance of contributing to the renewal of musical expression through science and technology. Conversely, specific problems related to contemporary composition have led to innovative, theoretical, methodological, and applied advances in the sciences with ramifications far beyond the world of music. The work carried out in the STMS joint research lab (Science and Technology of Music and Sound) is supported by the CNRS, Sorbonne Université and the French Ministry of Culture.
- NYU: New York University is a private research university based in New York City and founded in 1831 by Albert Gallatin as an institution to “admit based upon merit rather than birthright or social class”. NYU is the largest independent research university in the United States. The university has numerous research efforts, including founding the American Chemical Society and holding research partnerships with the Inception Institute of Artificial Intelligence and with major technology firms such as Twitter and IBM. The university has also since launched various internal research centers in the fields of artificial intelligence, history, culture, medicine, mathematics, philosophy, and economics.
- Georgia Tech: The Georgia Institute of Technology, also known as Georgia Tech, is a top-ranked public college and one of the leading research universities in the USA. Georgia Tech provides a technologically focused education to more than 25,000 undergraduate and graduate students in fields ranging from engineering, computing, and sciences, to business, design, and liberal arts. Georgia Tech’s wide variety of technologically-focused majors and minors consistently earn strong national rankings. Georgia Tech has six colleges and 28 schools focusing on Business, Computing, Design, Engineering, Liberal Arts, and Sciences.
- UC Berkeley: The University of California is a public research university. The University was founded in 1868, born out of a vision in the State Constitution of a university that would “contribute even more than California’s gold to the glory and happiness of advancing generations.” It is the oldest campus of the University of California system and ranks among the world’s top universities in major educational publications. Berkeley offers over 350 degree programs through its 14 colleges and schools.