Sedona.Biz – The Voice of Sedona and The Verde ValleySedona.Biz – The Voice of Sedona and The Verde Valley
    Sedona.Biz – The Voice of Sedona and The Verde Valley Sedona.Biz – The Voice of Sedona and The Verde Valley
    • Home
    • Sedona News
      • Arts and Entertainment
      • Bear Howard Chronicles
      • Business Profiles
      • City of Sedona
      • Elections
      • Goodies & Freebies
      • Mind & Body
      • Sedona News
    • Opinion
    • Real Estate
    • The Sedonan
    • Advertise
    • Sedona’s Best
    Sedona.Biz – The Voice of Sedona and The Verde ValleySedona.Biz – The Voice of Sedona and The Verde Valley
    Home»Editorials/Opinion»Opinion»Why pre-deployment testing and evaluation is not general AI safety
    Opinion

    Why pre-deployment testing and evaluation is not general AI safety

    January 1, 2025No Comments
    Facebook Twitter Pinterest LinkedIn Email Reddit WhatsApp
    shutterstock 2177929157
    Share
    Facebook Twitter LinkedIn Pinterest Email Reddit WhatsApp
    By David Stephen
    Sedona, AZ –The adoption of the pharmaceutical approach can be described as niche to AI safety. Pre-deployment testing and evaluation of OpenAI [o1], Anthropic [Claude 3.5 Sonnet] or others, only imply [circa] safety for those models, not general for AI.
    The misuses that are possible by AI tools across the internet [most times] may have nothing to do with OpenAI and Anthropic models. AI tools, for all kinds of purposes, are too numerous beyond what can be actively pre-tested or evaluated, by the US or UK AI safety institute.
    The pharmaceutical industry in most countries ensure that medications are approved, their ads are near accurate, the brands are identifiable among others [regulation] and there is the possibility for litigation or worse if they do not comply or mislead. There are several countries where counterfeit medications are available, but it is unlikely that a counterfeit medication group shows up in the mainstream.
    This is the problem with AI, where there are tools for all kinds of stuff available on first pages of search engines, on several social media, app stores and so forth.
    While the owners of the platforms have policies for safety, digital is also not the physical—without severe possibility for evasion and replication in short intervals. The question is how can AI be generally safe from synthetic [harmful, problematic] contents and several other misuses?
    How can AI also be understood for its mind? What are the components of AI that can be ascribed as the mind, then what are the features of those, to prospect how AI might be working and how to generally improve safety?
    General AI safety is not a case for just two major models, but around ways that AI can be safe across sources. If AI is a mind for digital contents or AI has a mind, what would be the components of that mind? Layers, nodes? If the components were extricated, how do they relay to define the outputs of AI? [Relays for the components of AI can be described as the math and compute that underpin AI]. Also, what are the channels from which alignment can be further introduced for general AI safety, across platforms? The human mind can be used to draw some parallels, while seeking out the mind for AI.
    There is a recent report by NIST, Pre-Deployment Evaluation of Anthropic’s Upgraded Claude 3.5 Sonnet, stating that, “The U.S. Artificial Intelligence Safety Institute (US AISI) and the UK Artificial Intelligence Safety Institute (UK AISI) conducted a joint pre-deployment evaluation of Anthropic’s latest model – the upgraded Claude 3.5 Sonnet (released October 22, 2024). US AISI and UK AISI ran separate but complementary tests to assess the model’s capabilities across four domains: (1) biological capabilities, (2) cyber capabilities, (3) software and AI development, and (4) safeguard efficacy. To assess the model’s relative capabilities and evaluate the potential real-world impacts of the upgraded Sonnet 3.5 across these four areas, US AISI and UK AISI compared its performance to a series of similar reference models: the prior version of Anthropic’s Sonnet 3.5, OpenAI’s o1-preview, and OpenAI’s GPT-4o.”
    There is a similar report by NIST, Pre-Deployment Evaluation of OpenAI’s o1 Model, stating that, “The U.S. Artificial Intelligence Safety Institute (US AISI) and the UK Artificial Intelligence Safety Institute (UK AISI) conducted a joint pre-deployment evaluation of OpenAI’s latest model, o1 (released December 5, 2024). US AISI and UK AISI conducted testing during a limited period of pre-deployment access to the o1 model. Testing was conducted by expert engineers, scientists, and subject matter specialists from staff at both Institutes, and the findings were shared with OpenAI before the model was publicly released. US AISI and UK AISI ran separate but complementary tests to assess the model’s capabilities across three domains: (1) cyber capabilities, (2) biological capabilities, (3) and software and AI development. To assess the model’s relative capabilities and evaluate the potential real-world impacts of o1 across these areas, US AISI and UK AISI compared its performance to a series of similar reference models: OpenAI’s o1-preview, OpenAI’s GPT-4o, and both the upgraded and earlier version of Anthropic’s Claude 3.5 Sonnet.”
    There is a recent announcement by OpenAI, Early access for safety testing, stating that, “We’re inviting safety researchers to apply for early access to our next frontier models. This early access program complements our existing frontier model testing process, which includes rigorous internal safety testing, external red teaming such as our Red Teaming Network⁠ and collaborations with third-party testing organizations, as well the U.S. AI Safety Institute and the UK AI Safety Institute. As models become more capable, we are hopeful that insights from the broader safety community can bring fresh perspectives, deepen our understanding of emerging risks, develop new evaluations, and highlight areas to advance safety research. As part of 12 Days of OpenAI⁠, we’re opening an application process for safety researchers to explore and surface the potential safety and security implications of the next frontier models.”

    Sedona Gift Shop

    Healing Paws

    This is an advertisement

    Comments are closed.


    What Would I Change?
    By Amaya Gayle Gregory

    What would I change if I could? You and I both know I can’t, but it’s a fun exercise anyway. I would have been less of a know-it-all on my spiritual journey. It seems to be a side-effect of the path. Spiritual folks develop an all-knowing buffer to protect against their inevitable surrender to the unknown, but understanding that now didn’t make it gentler on me or those I loved, let alone those that I deemed not capable of getting it 😉 Yeah … I’d have dropped the spiritual snob act. I’d have recognized that spiritual radicals are only different on the outside from radical right Christians, and that the surface doesn’t really matter as much as I thought. We are all doing our couldn’t be otherwise things, playing our perfect roles. I’d have learned to bow down humbly before my fellow man, regardless of whether I agreed with him or not. We’re all in this together and not one of us will get out alive. Read more→
    The Sedonan
    Need More Customers?
    Bear Howard Chronicles
    Humankind
    Tlaquepaque
    Verde Valley Wine Trail
    Recent Comments
    • Daniel J Sullivan MDJD on Innovative Affordable Workforce Housing for the City of Sedona
    • JOEY on Honoring Mom on Mother’s Day
    • Mary Allen on Innovative Affordable Workforce Housing for the City of Sedona
    • JB on 48 future RNs, 11 nursing bachelor’s degree earners feted during joyous Yavapai College pinning ceremony
    • Bill w on Innovative Affordable Workforce Housing for the City of Sedona
    • JB on Innovative Affordable Workforce Housing for the City of Sedona
    • styve on What Would I Change?
    • West Sedona Dave on Honoring Mom on Mother’s Day
    • Jill Dougherty on Innovative Affordable Workforce Housing for the City of Sedona
    • Bill w on Innovative Affordable Workforce Housing for the City of Sedona
    • JB on Honoring Mom on Mother’s Day
    • @Bill on Innovative Affordable Workforce Housing for the City of Sedona
    • Jill Dougherty on Innovative Affordable Workforce Housing for the City of Sedona
    • TJ Hall on Innovative Affordable Workforce Housing for the City of Sedona
    • Bill N. on Innovative Affordable Workforce Housing for the City of Sedona
    Archives

    What Would I Change?
    By Amaya Gayle Gregory

    What would I change if I could? You and I both know I can’t, but it’s a fun exercise anyway. I would have been less of a know-it-all on my spiritual journey. It seems to be a side-effect of the path. Spiritual folks develop an all-knowing buffer to protect against their inevitable surrender to the unknown, but understanding that now didn’t make it gentler on me or those I loved, let alone those that I deemed not capable of getting it 😉 Yeah … I’d have dropped the spiritual snob act. I’d have recognized that spiritual radicals are only different on the outside from radical right Christians, and that the surface doesn’t really matter as much as I thought. We are all doing our couldn’t be otherwise things, playing our perfect roles. I’d have learned to bow down humbly before my fellow man, regardless of whether I agreed with him or not. We’re all in this together and not one of us will get out alive. Read more→
    The Sedonan
    The Sedonan
    © 2025 All rights reserved. Sedona.biz.

    Type above and press Enter to search. Press Esc to cancel.