Sedona.Biz – The Voice of Sedona and The Verde ValleySedona.Biz – The Voice of Sedona and The Verde Valley
    Sedona.Biz – The Voice of Sedona and The Verde Valley Sedona.Biz – The Voice of Sedona and The Verde Valley
    • Home
    • Sedona
      • Arts and Entertainment
      • Bear Howard Chronicles
      • Business Profiles
      • City of Sedona
      • Elections
      • Goodies & Freebies
      • Mind & Body
      • Sedona News
    • Opinion
    • Real Estate
    • About
    • The Sedonan
    • Advertise
    • Sedona’s Best
    Sedona.Biz – The Voice of Sedona and The Verde ValleySedona.Biz – The Voice of Sedona and The Verde Valley
    Home»Editorials/Opinion»Opinion»Why pre-deployment testing and evaluation is not general AI safety
    Opinion

    Why pre-deployment testing and evaluation is not general AI safety

    January 1, 2025No Comments
    Facebook Twitter Pinterest LinkedIn Email Reddit WhatsApp
    shutterstock 2177929157
    Share
    Facebook Twitter LinkedIn Pinterest Email Reddit WhatsApp
    By David Stephen
    Sedona, AZ –The adoption of the pharmaceutical approach can be described as niche to AI safety. Pre-deployment testing and evaluation of OpenAI [o1], Anthropic [Claude 3.5 Sonnet] or others, only imply [circa] safety for those models, not general for AI.
    The misuses that are possible by AI tools across the internet [most times] may have nothing to do with OpenAI and Anthropic models. AI tools, for all kinds of purposes, are too numerous beyond what can be actively pre-tested or evaluated, by the US or UK AI safety institute.
    The pharmaceutical industry in most countries ensure that medications are approved, their ads are near accurate, the brands are identifiable among others [regulation] and there is the possibility for litigation or worse if they do not comply or mislead. There are several countries where counterfeit medications are available, but it is unlikely that a counterfeit medication group shows up in the mainstream.
    This is the problem with AI, where there are tools for all kinds of stuff available on first pages of search engines, on several social media, app stores and so forth.
    While the owners of the platforms have policies for safety, digital is also not the physical—without severe possibility for evasion and replication in short intervals. The question is how can AI be generally safe from synthetic [harmful, problematic] contents and several other misuses?
    How can AI also be understood for its mind? What are the components of AI that can be ascribed as the mind, then what are the features of those, to prospect how AI might be working and how to generally improve safety?
    General AI safety is not a case for just two major models, but around ways that AI can be safe across sources. If AI is a mind for digital contents or AI has a mind, what would be the components of that mind? Layers, nodes? If the components were extricated, how do they relay to define the outputs of AI? [Relays for the components of AI can be described as the math and compute that underpin AI]. Also, what are the channels from which alignment can be further introduced for general AI safety, across platforms? The human mind can be used to draw some parallels, while seeking out the mind for AI.
    There is a recent report by NIST, Pre-Deployment Evaluation of Anthropic’s Upgraded Claude 3.5 Sonnet, stating that, “The U.S. Artificial Intelligence Safety Institute (US AISI) and the UK Artificial Intelligence Safety Institute (UK AISI) conducted a joint pre-deployment evaluation of Anthropic’s latest model – the upgraded Claude 3.5 Sonnet (released October 22, 2024). US AISI and UK AISI ran separate but complementary tests to assess the model’s capabilities across four domains: (1) biological capabilities, (2) cyber capabilities, (3) software and AI development, and (4) safeguard efficacy. To assess the model’s relative capabilities and evaluate the potential real-world impacts of the upgraded Sonnet 3.5 across these four areas, US AISI and UK AISI compared its performance to a series of similar reference models: the prior version of Anthropic’s Sonnet 3.5, OpenAI’s o1-preview, and OpenAI’s GPT-4o.”
    There is a similar report by NIST, Pre-Deployment Evaluation of OpenAI’s o1 Model, stating that, “The U.S. Artificial Intelligence Safety Institute (US AISI) and the UK Artificial Intelligence Safety Institute (UK AISI) conducted a joint pre-deployment evaluation of OpenAI’s latest model, o1 (released December 5, 2024). US AISI and UK AISI conducted testing during a limited period of pre-deployment access to the o1 model. Testing was conducted by expert engineers, scientists, and subject matter specialists from staff at both Institutes, and the findings were shared with OpenAI before the model was publicly released. US AISI and UK AISI ran separate but complementary tests to assess the model’s capabilities across three domains: (1) cyber capabilities, (2) biological capabilities, (3) and software and AI development. To assess the model’s relative capabilities and evaluate the potential real-world impacts of o1 across these areas, US AISI and UK AISI compared its performance to a series of similar reference models: OpenAI’s o1-preview, OpenAI’s GPT-4o, and both the upgraded and earlier version of Anthropic’s Claude 3.5 Sonnet.”
    There is a recent announcement by OpenAI, Early access for safety testing, stating that, “We’re inviting safety researchers to apply for early access to our next frontier models. This early access program complements our existing frontier model testing process, which includes rigorous internal safety testing, external red teaming such as our Red Teaming Network⁠ and collaborations with third-party testing organizations, as well the U.S. AI Safety Institute and the UK AI Safety Institute. As models become more capable, we are hopeful that insights from the broader safety community can bring fresh perspectives, deepen our understanding of emerging risks, develop new evaluations, and highlight areas to advance safety research. As part of 12 Days of OpenAI⁠, we’re opening an application process for safety researchers to explore and surface the potential safety and security implications of the next frontier models.”

    Sedona Gift Shop

    Healing Paws

    This is an advertisement

    Comments are closed.

    Belief vs. Suspicion: Will Iran go the Way of Iraq, Gaza and Palestine?

    It was belief that Saddam had weapons of mass destruction that led us into war with Iraq. It’s suspicion that Iran seeks to build a nuclear warhead that has led us into war with them.
    As it turned out, Saddam had no weapons of mass destruction. Will it be the same with Iran?

    Read more→

    The Sedonan
    Need More Customers?
    Bear Howard Chronicles
    Humankind
    Tlaquepaque
    Verde Valley Wine Trail
    Recent Comments
    • TJ Hall on Belief vs. Suspicion: Will Iran go the Way of Iraq, Gaza and Palestine?
    • Michael Johnson on Belief vs. Suspicion: Will Iran go the Way of Iraq, Gaza and Palestine?
    • JB on Belief vs. Suspicion: Will Iran go the Way of Iraq, Gaza and Palestine?
    • floyd gardner on Belief vs. Suspicion: Will Iran go the Way of Iraq, Gaza and Palestine?
    • Skip Daum on Rowe Fine Art Gallery Salutes Free Spirits
    • JB on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • TJ Hall on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • Jill Dougherty on The Rise of the Enforcement Class
    • Jill Dougherty on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • Rob Schwab on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • Jill Dougherty on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • TJ Hall on The Rise of the Enforcement Class
    • JB on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • @mkJeeves on Cottonwood, Verde Valley Residents Join Largest Protest Yet to Reject Abuses of Power
    • TJ Hall on The Rise of the Enforcement Class
    Archives
    The Sedonan
    © 2025 All rights reserved. Sedona.biz.

    Type above and press Enter to search. Press Esc to cancel.