VANCOUVER, British Columbia--(BUSINESS WIRE)--Nov 28, 2024--
As the holiday season draws near and friends and families gather, Purdys Chocolatier is sending a sweet reminder to Canadians: No matter how full you feel after a hearty holiday dinner, there’s always room for Purdys. The Canadian chocolatier is unveiling a heartwarming, yet humorous new commercial, capturing the universal truth of sweet indulgence during the festive season.
This press release features multimedia. View the full release here: https://www.businesswire.com/news/home/20241128233812/en/
A Message as Sweet as Chocolate - Watch the commercialhere
In the brand spot, three households celebrate the holidays as many do: decked out in cozy sweaters, festive paper crowns, and surrounded by the people nearest and dearest to them. The commercial transports viewers to nostalgic settings filled with turkey, mashed potatoes, hot pot, and the overwhelming feeling of being pleasantly stuffed and ready for a well-earned nap. The scenes reveal how different people truly believe they’re completely full - until a box of Hedgehogs from Purdys Chocolatier appears on the table.
“It’s a feeling we can all relate to,” explains Julia Cho, Director of Marketing at Purdys Chocolatier. “The moment you see Purdys chocolate, something magical happens and there’s always room for something sweet, even after the most epic holiday meal!”
The commercial ends with mom at the head of the table taking her first bite of a Purdys Hedgehog, reminding viewers of the hilarious yet relatable realization that no one is ever truly too full for dessert. In fact, in a survey recently conducted by Leger, over two-thirds of Canadians agree that they always manage to find room for “something sweet” at the end of a holiday dinner.
“At Purdys Chocolatier, we believe the holidays are about savouring the simple moments that bring us together — and there’s nothing like enjoying a box of Purdys chocolates with your loved ones,” shares Cho.
What to Bring to the Table
This year’s irresistible holiday collection is filled with new chocolates, as well as fan favourites packaged in festive wrapping. Those looking to bring something new to their holiday gatherings can pick up: White Himalayan Pink Salt Caramels, Snowballs, Orange Slices, the Vegan Dark Chocolate Collection, and the Nutty & Chewy Collection. Those looking to bring a classic, yet festive treat can pick up: Hedgehogs, the Festive Wreath Gift Box, Peppermint Bark, and more. To view the entire Purdys Chocolatier Christmas collection, visit https://www.purdys.com/holiday-gift-guide.
The spot will appear on YouTube, Meta, TikTok and a TV campaign across Canada from November 28 until December 23.
Head to www.purdys.com to shop online or visit a Purdys Chocolatier shop near you.
About Purdys Chocolatier
Purdys Chocolatier has been Canada's Chocolatier since 1907. As a proud, family-owned company, they make a wide range of chocolate gifts from premium ingredients and sustainable cocoa. Community and people are at the heart of Purdys Chocolatier, and together with their Purple Partnerships initiative and Seasonal Fundraising Program, they find opportunities to support, advocate and craft more kindness.
For more information about Purdys Chocolatier, visit https://www.purdys.com.
(Photo: Business Wire)
WASHINGTON (AP) — With artificial intelligence at a pivotal moment of development, the federal government is about to transition from one that prioritized AI safeguards to one more focused on eliminating red tape.
That’s a promising prospect for some investors but creates uncertainty about the future of any guardrails on the technology, especially around the use of AI deepfakes in elections and political campaigns.
President-elect Donald Trump has pledged to rescind President Joe Biden’s sweeping AI executive order, which sought to protect people’s rights and safety without stifling innovation. He hasn’t specified what he would do in its place, but the platform of the Republican National Committee, which he recently reshaped, said AI development should be “rooted in Free Speech and Human Flourishing.”
It’s an open question whether Congress, soon to be fully controlled by Republicans, will be interested in passing any AI-related legislation. Interviews with a dozen lawmakers and industry experts reveal there is still interest in boosting the technology’s use in national security and cracking down on non-consensual explicit images.
Yet the use of AI in elections and in spreading misinformation is likely to take a backseat as GOP lawmakers turn away from anything they view as potentially suppressing innovation or free speech.
“AI has incredible potential to enhance human productivity and positively benefit our economy,” said Rep. Jay Obernolte, a California Republican widely seen as a leader in the evolving technology. “We need to strike an appropriate balance between putting in place the framework to prevent the harmful things from happening while at the same time enabling innovation.”
Artificial intelligence interests have been expecting sweeping federal legislation for years. But Congress, gridlocked on nearly every issue, failed to pass any artificial intelligence bill, instead producing only a series of proposals and reports.
Some lawmakers believe there is enough bipartisan interest around some AI-related issues to get a bill passed.
“I find there are Republicans that are very interested in this topic,” said Democratic Sen. Gary Peters, singling out national security as one area of potential agreement. "I am confident I will be able to work with them as I have in the past.”
It’s still unclear how much Republicans want the federal government to intervene in AI development. Few showed interest before this year's election in regulating how the Federal Election Commission or the Federal Communications Commission handled AI-generated content, worrying that it would raise First Amendment issues at the same time that Trump’s campaign and other Republicans were using the technology to create political memes.
The FCC was in the middle of a lengthy process for developing AI-related regulations when Trump won the presidency. That work has since been halted under long-established rules covering a change in administrations.
Trump has expressed both interest and skepticism in artificial intelligence.
During a Fox Business interview earlier this year, he called the technology “very dangerous” and “so scary” because “there’s no real solution.” But his campaign and supporters also embraced AI-generated images more than their Democratic opponents. They often used them in social media posts that weren’t meant to mislead, but rather to further entrench Republican political views.
Elon Musk, Trump's close adviser and a founder of several companies that rely on AI, also has shown a mix of concern and excitement about the technology, depending on how it is applied.
Musk used X, the social media platform he owns, to promote AI-generated images and videos throughout the election. Operatives from Americans for Responsible Innovation, a nonprofit focused on artificial intelligence, have publicly been pushing Trump to tap Musk as his top adviser on the technology.
“We think that Elon has a pretty sophisticated understating of both the opportunities and risks of advanced AI systems,” said Doug Calidas, a top operative from the group.
But Musk advising Trump on artificial intelligence worries others. Peters argued it could undercut the president.
“It is a concern,” said the Michigan Democrat. “Whenever you have anybody that has a strong financial interest in a particular technology, you should take their advice and counsel with a grain of salt.”
In the run-up to the election, many AI experts expressed concern about an eleventh-hour deepfake — a lifelike AI image, video or audio clip — that would sway or confuse voters as they headed to the polls. While those fears were never realized, AI still played a role in the election, said Vivian Schiller, executive director of Aspen Digital, part of the nonpartisan Aspen Institute think tank.
“I would not use the term that I hear a lot of people using, which is it was the dog that didn’t bark,” she said of AI in the 2024 election. “It was there, just not in the way that we expected.”
Campaigns used AI in algorithms to target messages to voters. AI-generated memes, though not lifelike enough to be mistaken as real, felt true enough to deepen partisan divisions.
A political consultant mimicked Joe Biden’s voice in robocalls that could have dissuaded voters from coming to the polls during New Hampshire's primary if they hadn’t been caught quickly. And foreign actors used AI tools to create and automate fake online profiles and websites that spread disinformation to a U.S. audience.
Even if AI didn’t ultimately influence the election outcome, the technology made political inroads and contributed to an environment where U.S. voters don’t feel confident that what they are seeing is true. That dynamic is part of the reason some in the AI industry want to see regulations that establish guidelines.
“President Trump and people on his team have said they don’t want to stifle the technology and they do want to support its development, so that is welcome news,” said Craig Albright, the top lobbyist and senior vice president at The Software Alliance, a trade group whose members include OpenAI, Oracle and IBM. “It is our view that passing national laws to set the rules of the road will be good for developing markets for the technology."
AI safety advocates during a recent meeting in San Francisco made similar arguments, according to Suresh Venkatasubramanian, director of the Center for Tech Responsibility at Brown University.
“By putting literal guardrails, lanes, road rules, we were able to get cars that could roll a lot faster,” said Venkatasubramanian, a former Biden administration official who helped craft White House principles for approaching AI.
Rob Weissman, co-president of the advocacy group Public Citizen, said he's not hopeful about the prospects for federal legislation and is concerned about Trump’s pledge to rescind Biden’s executive order, which created an initial set of national standards for the industry. His group has advocated for federal regulation of generative AI in elections.
“The safeguards are themselves ways to promote innovation so that we have AI that’s useful and safe and doesn’t exclude people and promotes the technology in ways that serve the public interest,” he said.
The Associated Press receives support from several private foundations to enhance its coverage of elections and democracy, and from the Omidyar Network to support coverage of artificial intelligence and its impact on society. AP is solely responsible for all content. See more about AP’s democracy initiative here and a list of supporters and funded coverage areas at AP.org
FILE - President-elect Donald Trump speaks during a meeting with the House GOP conference, Nov. 13, 2024, in Washington. (Allison Robbert/Pool via AP, File)
Elizabeth Kelly, Director of the U.S. AI Safety Institute, listens to speakers at the convening of the International Network of AI Safety Institutes at the Golden Gate Club at the Presidio in San Francisco, Wednesday, Nov. 20, 2024. (AP Photo/Jeff Chiu)