The company wants manual auditors to pass the AI application
安全到达彼岸依
发表于 2023-10-26 07:57:16
1327
0
0
Enterprises weighing the risks and benefits of generative artificial intelligence (AI) technology are facing the challenge that social media platforms have long been striving to address: preventing technology from being maliciously exploited.
Drawing on the experience of these platforms, business technology leaders are starting to combine software based "guardrails" with manual auditors to limit their use within prescribed limits.
AI models such as OpenAI's GPT-4 have been trained through extensive internet content. With the correct prompts, large language models can generate a large amount of toxic content inspired by the darkest corners of the internet. This means that content auditing needs to occur at the source (i.e., when AI models are trained) and their heavily generated outputs.
TurboTax software developer Intuit Inc. (INTU), headquartered in Mountain View, California, recently released Intuit Assist, a generative AI based assistant that provides financial advice to customers. At present, this assistant is only available for a limited number of users, relying on large language models trained on internet data and fine-tuning models based on Intuit's own data.
Intuit Chief Information Security Officer Atticus Tysen
The company's Chief Information Security Officer, Atticus Tysen, stated that the company is currently planning to form a team of eight full-time auditors to review the content entering and exiting this large language model driven system, including helping to prevent employees from leaking sensitive company data.
Tysen said, "When we try to provide truly meaningful and specific answers around financial issues, we don't know how effective these models are. So for us, adding manpower to this loop is very important
Tysen stated that Intuit's self-developed content review system uses a separate large language model to automatically label content it deems offensive, such as profanity, and is currently in its early stages. He said, for example, if a customer asks questions unrelated to financial guidelines or attempts to launch a prompt injection attack, the customer will also be automatically banned by the system. These attacks may include inducing chatbots to disclose customer data or the way they operate.
Then, the manual auditor will be reminded to review the text and be able to send it to the model building team, thereby improving the system's ability to block or identify harmful content. If Intuit's customers believe that their prompt words have been incorrectly marked, or if they believe that the AI assistant has generated inappropriate content, they can also notify the company.
Although there is currently no company specializing in AI content auditing, Intuit is supplementing its workforce with contractors trained in social media content auditing. Like the so-called prompt word engineer, AI content reviewers may become part of a new type of job opportunity created by AI.
Tysen said that the ultimate goal of Caijie Group is to have its AI audit model complete most of the content audit work for its AI assistants, thereby reducing the harmful content that humans may come into contact with. But he said that at present, generative AI is not enough to completely replace manual auditors.
Social media companies such as Facebook and Instagram's parent company Meta have long relied on outsourced human auditors to review and filter offensive posts on their platforms, providing best practices and warnings for the future development path of AI auditing.
In recent years, AI companies like OpenAI have recruited personnel to review and classify harmful text obtained online and generated by AI itself. These classified paragraphs are used to create AI security filters for ChatGPT, preventing users of the chat robot from accessing similar content.
OpenAI also collaborated with its partner and biggest supporter Microsoft to develop what Microsoft calls Azure AI Content Safety service, which utilizes AI to automatically detect unsafe images and text, including hate, violence, sexual, and self harm content. Microsoft is using its security services to prevent harmful content from appearing in its generative AI tools, including GitHub Copilot and Copilot for Office software.
Eric Boyd, Vice President of Microsoft AI Platform Enterprise, said, "These AI systems are indeed quite powerful. With the right instructions, they can do various things
Other technology leaders are studying the potential of manual auditing or investing in third-party software like Microsoft. Analysts say that content security filters will soon become a necessary condition for businesses to register and use generative AI tools sold by any supplier.
Syneos Health, the Chief Information and Digital Officer of Syneos Health, a biopharmaceutical services company located in Morrisville, North Carolina, has stated that the company will consider hiring content auditors at some point next year. During this period, the training data used by the AI model will be reviewed one by one through manual feedback.
Pickett said, "We will do this in a precise surgical manner, but in a broader sense, some level of review and supervision has many benefits
Forrester analyst Brandon Purcell focuses on responsible and ethical AI usage issues, stating that people are increasingly interested in "responsible AI", with the aim of making AI algorithms more transparent or auditable and reducing the unintended negative consequences of AI.
He said, "Everyone is interested in this because they realize that if they don't do it well, they will face reputational, regulatory, and revenue risks
CandyLake.com 系信息发布平台,仅提供信息存储空间服务。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
猜你喜欢
- Domestic CAR-T enterprise Legendary Biology's first quarter losses narrowed, and the company is expected to achieve profitability in 2026
- Taobao launches live streaming full hosting service for corporate CEOs
- Baidu's big model ERNIE 4.0 Turbo is fully open to enterprise users
- Google reportedly plans to acquire cybersecurity startup Wiz for $23 billion, making it the largest acquisition in its history
- Luckin Coffee and 6 other coffee companies were interviewed
- Kanye's shopping receipts have been exposed! Miniso is suspected of leaking privacy, and the company has responded
- Several companies, including Jike and Tesla, have responded by denying illegal surveying and mapping under the pretext of intelligent driving of automobiles
- Is a well-known eVTOL company on the brink of bankruptcy?
- AI Agents: A New Track for Technology Companies to Compete in
- Baidu: Within 3 days of its release, over 5000 companies have queued up to apply for testing under the name 'Second Da'
-
アリババは、26億5000万ドルのドル建て優先無担保手形と170億元の人民元建て優先無担保手形の定価を発表した。ドル債の発行は2024年11月26日に終了する予定です。人民元債券の発行は2024年11月28日に終了する予定だ ...
- SOGO
- 前天 09:05
- 支持
- 反对
- 回复
- 收藏
-
スターバックスが中国事業の株式売却の可能性を検討していることが明らかになった。 11月21日、外国メディアによると、スターバックスは中国事業の株式売却を検討している。関係者によると、スターバックスは中国事 ...
- 献世八宝掌
- 昨天 16:29
- 支持
- 反对
- 回复
- 收藏
-
【意法半導体CEO:中国市場は非常に重要で華虹と協力を展開】北京時間11月21日、意法半導体(STM.N)は投資家活動の現場で、同社が中国ウェハー代工場の華虹公司(688347.SH)と協力していると発表した。伊仏半導体 ...
- 黄俊琼
- 昨天 14:29
- 支持
- 反对
- 回复
- 收藏
-
【ナスダック中国金龍指数は1%下落した。人気の中概株の多くは下落した】現地時間11月21日、ナスダック中国金龍指数は1%下落し、人気の中概株の多くは下落し、必死に10%超下落し、愛奇芸は7%超下落し、百度は6%近く ...
- 比尔992
- 4 小时前
- 支持
- 反对
- 回复
- 收藏