{"id":168920,"date":"2023-08-02T14:34:29","date_gmt":"2023-08-02T19:34:29","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2023\/08\/meta-open-sources-framework-for-generating-sounds-and-music"},"modified":"2023-08-02T14:34:29","modified_gmt":"2023-08-02T19:34:29","slug":"meta-open-sources-framework-for-generating-sounds-and-music","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2023\/08\/meta-open-sources-framework-for-generating-sounds-and-music","title":{"rendered":"Meta open sources framework for generating sounds and music"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/meta-open-sources-framework-for-generating-sounds-and-music2.jpg\"><\/a><\/p>\n<p>The day is fast approaching when generative AI won\u2019t only write and create images in a convincingly human-like style, but compose music and sounds that pass for a professional\u2019s work, too. This morning, Meta announced AudioCraft, a framework to generate what it describes as \u201chigh-quality,\u201d \u201crealistic\u201d audio and music from short text descriptions, or prompts. It\u2019s not Meta\u2019s first foray into audio generation \u2014 the tech giant open sourced an AI-powered music generator, MusicGen, in June \u2014 but Meta claims that it\u2019s made advances that vastly improve the quality of AI-generated sounds, such as dogs barking, cars honking and footsteps on a wooden floor.<\/p>\n<p>In a blog post shared with TechCrunch, Meta\u2026 More.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The day is fast approaching when generative AI won\u2019t only write and create images in a convincingly human-like style, but compose music and sounds that pass for a professional\u2019s work, too. This morning, Meta announced AudioCraft, a framework to generate what it describes as \u201chigh-quality,\u201d \u201crealistic\u201d audio and music from short text descriptions, or prompts. [\u2026]<\/p>\n","protected":false},"author":578,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[42,6,1491],"tags":[],"class_list":["post-168920","post","type-post","status-publish","format-standard","hentry","category-media-arts","category-robotics-ai","category-transportation"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/168920","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/578"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=168920"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/168920\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=168920"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=168920"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=168920"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}