ContentSproute

us-technology

A ChatGPT ‘router’ that automatically selects the right OpenAI model for your job appears imminent

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now In the 2.5 years since OpenAI debuted ChatGPT, the number of large language models (LLMs) that the company has made available as options to power its hit chatbot has steadily grown. In fact, there are now a total of 7 (!!!) different AI models that paying ChatGPT subscribers (of the $20 Plus tier and more expensive tiers) can choose between when interacting with the trusty chatbot — each with its own strengths and weaknesses. But how should a user decide which one to use for their particular prompt, question, or task? After all, you can only pick one at a time. The AI Impact Series Returns to San Francisco – August 5 The next phase of AI is here – are you ready? Join leaders from Block, GSK, and SAP for an exclusive look at how autonomous agents are reshaping enterprise workflows – from real-time decision-making to end-to-end automation. Secure your spot now – space is limited: https://bit.ly/3GuuPLF Is help on the way? Help appears to be on the way imminently from OpenAI — as reports emerged over the last few days on X from AI influencers, including OpenAI’s own researcher “Roon (@tszzl on X)” (speculated to be technical team member Tarun Gogineni) — of a new “router” function that will automatically select the best OpenAI model to respond to the user’s input on the fly, depending on the specific input’s content. As Roon posted on the social network X yesterday, July 20, 2025, in since-deleted response to influencer Lisan al Gaib’s statement that they “don’t want a model router I want to be able to select the models I use”: “You’ll still be able to select. This is a product to make sure that doctors aren’t stuck on 4o-mini” Similarly, Yuchen Jin, Co-founder & CTO of AI inference cloud provider Hyperbolic Labs, wrote in an X post on July 19. “Heard GPT-5 is imminent, from a little bird. It’s not one model, but multiple models. It has a router that switches between reasoning, non-reasoning, and tool-using models. That’s why Sam said they’d “fix model naming”: prompts will just auto-route to the right model. GPT-6 is in training. I just hope they’re not delaying it for more safety tests. 🙂“ While a presumably far more advanced GPT-5 model would (and will) be huge news if and when released, the router may make life much easier and more intelligent for the average ChatGPT subscriber. It would also follow on the heels of other third-party products such as the web-based Token Monster chatbot, which automatically select and combine responses from multiple third-party LLMs to respond to user queries. Asked about the router idea and comments from “Roon,” an OpenAI spokesperson declined to provide a response or further information at this time. Solving the overabundance of choice problem To be clear, every time OpenAI has released a new LLM to the public, it has diligently shared in either a blog post or release notes or both what it thinks that particular model is good for and designed to help with. For example, OpenAI’s “o” series reasoning models — o3, o4-mini, o4-mini high — have performed better on math, science, and coding tests thanks to benchmarking tests, while non-reasoning models like the new GPT-4.5 and 4.1 seem to do better at creative writing and communications tasks. Dedicated AI influencers and power users may understand very well what all these different models are good and not so good at. But regular users who don’t follow the industry as closely, nor have the time and finances available to test them all out on the same input prompts and compare the outputs, will understandably struggle to make sense of the bewildering array of options. That could mean they’re missing out on smarter, more intelligent, or more capable responses from ChatGPT for their task at hand. And in the case of fields like medicine, as Roon alluded to, the difference could be one of life or death. It’s also interesting to speculate on how an automatic LLM router might change public perceptions toward and adoption of AI more broadly. ChatGPT already counted 500 million active users as of March. If more of these people were automatically guided toward more intelligent and capable LLMs to handle their AI queries, the impact of AI on their workloads and that of the entire global economy would seem likely to be felt far more acutely, creating a positive “snowball” effect. That is, as more people saw more gains from ChatGPT automatically choosing the right AI model for their queries, and as more enterprises reaped greater efficiency from this process, more and more individuals and organizations would likely be convinced by the utility of AI and be more willing to pay for it, and as they did so, even more AI-powered workflows would spread out in the world. But right now, this is presumably all being held back a little by the fact that the ChatGPT model picker requires the user to A. know they even have a choice of models and B. have some level of informed awareness of what these models are good for. It’s all still a manually driven process. Like going to the supermarket in your town and staring at aisles of cereal and different sauces, the average ChatGPT user is currently faced with an overabundance of choice. Hopefully any hypothetical OpenAI router seamlessly helps direct them to the right model product for their needs, when they need it — like a trusty shopkeeper showing up to free you from your product paralysis. Daily insights on business use cases with VB Daily If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

A ChatGPT ‘router’ that automatically selects the right OpenAI model for your job appears imminent Read More »

How to Migrate from OpenAI to Cerebrium for Cost-Predictable AI Inference

If you’re building an AI application, you probably started with OpenAI’s convenient APIs. However, as your application scales, you’ll need more control over costs, models, and infrastructure. Cerebrium is a serverless AI infrastructure platform that lets you run open-source models on dedicated hardware with predictable, time-based pricing instead of token-based billing. This guide will show you how to build a complete chat application with OpenAI, migrate it to Cerebrium by changing just two lines of code, and add performance and cost tracking to compare the two approaches to AI inference using real data. When you’re done, you’ll have a working chat application that demonstrates the practical differences between token-based and compute-based pricing models, and the insights you need to choose the right approach for your use case. Prerequisites To follow along with this guide, you’ll need Python 3.10 or higher installed on your system. You’ll also need the following (all free): OpenAI API key. Cerebrium account (includes free tier access to test GPU instances up to A10 level). Hugging Face token (free account required). Llama 3.1 model access on Hugging Face. Visit meta-llama/Meta-Llama-3.1-8B-Instruct and click “Request access” to get approval from Meta (typically takes a few minutes to a few hours). Familiarity with Python and API calls is helpful, but we’ll walk through each step in detail. Creating an OpenAI Chatbot We’ll build a complete chat application that works with OpenAI as our foundation and enhance it throughout the tutorial without ever needing to modify the core chat logic. Create a new directory for the project and set up the basic structure: mkdir openai-cerebrium-migration cd openai-cerebrium-migration Install the dependencies: pip install openai==1.55.0 python-dotenv==1.0.0 art==6.1 colorama==0.4.6 Create a .env file to store API credentials: OPENAI_API_KEY=your_openai_api_key_here CEREBRIUM_API_KEY=your_cerebrium_api_key_here CEREBRIUM_ENDPOINT_URL=your_cerebrium_endpoint_url_here Replace your_openai_api_key_here with your actual OpenAI API key. Now we’ll build the chat.py file step by step. Start by creating the file and adding the imports: import os import time from dotenv import load_dotenv from openai import OpenAI from art import text2art from colorama import init, Fore, Style These imports handle environment variables, OpenAI client creation, ASCII art generation, and colored terminal output. Add the initialization below the imports: load_dotenv() init(autoreset=True) Add this display_intro function: def display_intro(use_cerebrium, endpoint_name): print(“n”) if use_cerebrium: ascii_art = text2art(“Cerebrium”, font=”tarty1″) print(f”{Fore.MAGENTA}{ascii_art}{Style.RESET_ALL}”) else: ascii_art = text2art(“OpenAI”, font=”tarty1″) print(f”{Fore.WHITE}{ascii_art}{Style.RESET_ALL}”) print(f”Connected to: {Fore.CYAN}{endpoint_name}{Style.RESET_ALL}”) print(“nType ‘quit’ or ‘exit’ to end the chatn”) This function provides visual feedback when we switch between endpoints. Add the main function that handles the chat logic: def main(): # OpenAI endpoint client = OpenAI(api_key=os.getenv(“OPENAI_API_KEY”)) model = “gpt-4o-mini” endpoint_name = “OpenAI (GPT-4o-mini)” use_cerebrium = False display_intro(use_cerebrium, endpoint_name) conversation = [] while True: user_input = input(“You: “).strip() if user_input.lower() in [‘quit’, ‘exit’, ‘bye’]: print(“Goodbye!”) break if not user_input: continue conversation.append({“role”: “user”, “content”: user_input}) This function sets up the endpoint configuration and handles the basic chat loop. Add the response handling logic inside the main function’s while loop: try: print(“Bot: “, end=””, flush=True) chat_completion = client.chat.completions.create( messages=conversation, model=model, stream=True, stream_options={“include_usage”: True}, temperature=0.7 ) bot_response = “” for chunk in chat_completion: if chunk.choices[0].delta.content: content = chunk.choices[0].delta.content print(content, end=””, flush=True) bot_response += content print() conversation.append({“role”: “assistant”, “content”: bot_response}) except Exception as e: print(f”❌ Error: {e}”) conversation.pop() Finally, add the script execution guard at the end of the file: if __name__ == “__main__”: main() Test the chatbot by running: You’ll see the OpenAI ASCII art, and you can start chatting with GPT-4o mini. Ask a question to verify that the app works correctly. Responses will stream in real-time. Deploying a Cerebrium Endpoint With vLLM and Llama 3.1 Now we’ll create a Cerebrium endpoint that serves the same OpenAI-compatible interface using vLLM and an open-source model. When we’re done, we’ll be able to switch to a self-hosted open-source model endpoint by changing just two lines of code. Configuring Hugging Face Access for Llama 3.1 First, make sure you have access to the Llama 3.1 model on Hugging Face. If you haven’t already requested access, visit meta-llama/Meta-Llama-3.1-8B-Instruct and click “Request access”. Next, create a Hugging Face token by going to Hugging Face settings, clicking “New token”, and selecting “Read” permissions. Add your Hugging Face token to your Cerebrium project secrets. Go to your Cerebrium dashboard, select your project, and add HF_AUTH_TOKEN with your Hugging Face token as the value. Setting Up a Cerebrium Account and API Access Create a free Cerebrium account and navigate to your dashboard. In the “API Keys” section, copy your session token and save it for later – you’ll need it to authenticate with the deployed endpoint. Add the session token to the .env file as a CEREBRIUM_API_KEY variable: OPENAI_API_KEY=your_openai_api_key_here CEREBRIUM_API_KEY=your_cerebrium_api_key_here CEREBRIUM_ENDPOINT_URL=your_cerebrium_endpoint_url_here Building the OpenAI-Compatible vLLM Endpoint Start by installing the Cerebrium CLI and creating a new project: pip install cerebrium cerebrium login cerebrium init openai-compatible-endpoint cd openai-compatible-endpoint We’ll build the main.py file step by step to understand each component. Start with the imports and authentication: from vllm import SamplingParams, AsyncLLMEngine from vllm.engine.arg_utils import AsyncEngineArgs from pydantic import BaseModel from typing import Any, List, Optional, Union, Dict import time import json import os from huggingface_hub import login login(token=os.environ.get(“HF_AUTH_TOKEN”)) These imports provide the vLLM async engine for model inference, Pydantic models for data validation, and Hugging Face authentication for model access. Add the vLLM engine configuration: engine_args = AsyncEngineArgs( model=”meta-llama/Meta-Llama-3.1-8B-Instruct”, gpu_memory_utilization=0.9, # Set GPU memory utilization max_model_len=8192 # Set max model length ) engine = AsyncLLMEngine.from_engine_args(engine_args) This configuration uses 90% of available GPU memory and sets an 8K-token context window, optimizing for throughput while maintaining reasonable memory usage. Now add the Pydantic models that define the OpenAI-compatible response format: class Message(BaseModel): role: str content: str class ChoiceDelta(BaseModel): content: Optional[str] = None function_call: Optional[Any] = None refusal: Optional[Any] = None role: Optional[str] = None tool_calls: Optional[Any] = None class Choice(BaseModel): delta: ChoiceDelta finish_reason: Optional[str] = None index: int logprobs: Optional[Any] = None class Usage(BaseModel): completion_tokens: int = 0 prompt_tokens: int = 0 total_tokens: int = 0 class ChatCompletionResponse(BaseModel): id: str object: str created: int model: str choices: List[Choice] service_tier: Optional[str] = “default” system_fingerprint: Optional[str] = “fp_cerebrium_vllm” usage: Optional[Usage] = None These models ensure

How to Migrate from OpenAI to Cerebrium for Cost-Predictable AI Inference Read More »

Kapa.ai (YC S23) is hiring a software engineers (EU remote)

Create enterprise-grade AI assistants from your content Software Engineer (Full-stack) $100K – $150K / 0.10% – 0.30% Location GB / EG / RU / UA / TR / FR / IT / ES / PL / RO / KZ / NL / BE / SE / CZ / GR / PT / HU / AT / CH / BG / DK / FI / NO / SK / LT / EE / DE / Remote (GB; EG; RU; UA; TR; FR; IT; ES; PL; RO; KZ; NL; BE; SE; CZ; GR; PT; HU; AT; CH; BG; DK; FI; NO; SK; LT; EE; DE) Visa US citizenship/visa not required Connect directly with founders of the best YC-funded startups. Apply to role › About the role As a software engineer you will work across the stack on the Kapa systems that answer thousands of developer questions per day. Check out Docker’s documentation for a live example of what kapa is. In this role, you will: Work directly with the founding team and our research engineers. Scale the infrastructure that powers the Kapa RAG engine (Python). Experiment with new features in the Kapa analytics platform (React + Python). Work on the client integrations which are used to deploy Kapa for our customers (React + Python). Give Kapa access to new kinds of data (Python). Maintain our React SDK. You may be a good fit if you have: A degree in computer science, machine learning, mathematics, statistics or a related field. 3+ years of software engineering experience working on complex systems in both backend and frontend. An affinity for machine learning, deep learning (including LLMs) and natural language processing. The ability to work effectively in a fast in a environment where things are sometimes loosely defined. * This is neither an exhaustive nor necessary set of attributes. Even if none of these apply to you, but you believe you will contribute to kapa.ai, please reach out. About kapa.ai kapa.ai makes it easy for technical companies to build AI support and onboarding bots for their users. Teams at +150 leading startups and enterprises incl. OpenAI, Mixpanel, Mapbox, Docker, Next.js and Prisma use kapa to level up their developer experience and reduce support. We enable companies to use their existing technical knowledge sources incl. docs, tutorials, chat logs, and GitHub issues to generate AI bots that answers developer questions automatically. More than 750k developers have access to kapa.ai via website widgets, Slack/Discord bots, API integrations, or via Zendesk. We’ve been fortunate to be funded by some of the greatest AI investors in Silicon Valley: Initialized Capital (Garry Tan, Alexis Ohanian), Y Combinator, Amjad Masad and Michele Catasta (Replit), and Douwe Kiela (RAG paper author and founder of Contextual AI), and other folks incl. angels at OpenAI. Founded:2023 Batch:S23 Team Size:14 Status:Active Founders Read More

Kapa.ai (YC S23) is hiring a software engineers (EU remote) Read More »

Complete silence is always hallucinated as “ترجمة نانسي قنقر” in Arabic

Comment options {{title}} VAD, probably. I’ve only tried the turbo one, but what I can say is that v3 is different from the earlier models. It looks like it doesn’t have the audio descriptions to fall back on and produces hallucinations instead. The earlier models will also produce some miscellaneous crap when they encounter silence (they do this regardless of language), but there are more options for how to deal with that. For example, these things can be effective for the small model (but not for v3): the suppress_tokens trick setting initial prompt to something like “.” adjusting logprob_threshold to -0.4 (works for this empty audio, probably not good for general use) You must be logged in to vote 0 replies Comment options {{title}} is there any good arabic model you guys found which is better than large v3 ? @misutoneko @puthre You must be logged in to vote 1 reply Comment options {{title}} Voxtral was released a few days ago and looks promising Comment options {{title}} I found a similar thing happens in German where it says “Untertitelung des ZDF für funk, 2017.” For both German and Arabic I found that this pretty much only happens at the very end of videos / when there is sustained silence. You must be logged in to vote 0 replies Comment options {{title}} Essentially this seems to be an artifact of the fact that Whisper was trained on (amongst other things) YouTube audio + available subtitles. Often subtitlers add their copyright notice onto the end of the subtitles, and the end of the videos are often credits with music, applause, or silence. Thus whisper learned that silence == “copyright notice”. See some research for the Norwegian example here: https://medium.com/@lehandreassen/who-is-nicolai-winther-985409568201 You must be logged in to vote 0 replies Comment options {{title}} In English there is always applause You must be logged in to vote 0 replies Comment options {{title}} this also happens when you don’t speak into the voice mode, the transcript usually results in the same Arabic phrase You must be logged in to vote 0 replies Comment options {{title}} I’ve also seen this happen a lot in English with Skyeye: It also happens a lot with hallucinations saying stuff like “This is the end of the video, remember to like and subscribe” You must be logged in to vote 0 replies Comment options {{title}} I have built https://arabicworksheet.com for arabic learning from absolute beginners to professional speakers. It created dynamic exercises and worksheets based on your level and topics. Behind the scene I have used Gemini 2.5-pro & GPT-4o for overall agentic workflows. You must be logged in to vote 1 reply Comment options {{title}} Ok? This doesn’t have anything to do with the topic of this discussion Comment options {{title}} In german it’s “Vielen Dank” (Thank you very much) You must be logged in to vote 0 replies Comment options {{title}} You must be logged in to vote 0 replies Read More

Complete silence is always hallucinated as “ترجمة نانسي قنقر” in Arabic Read More »

We have made the decision to not continue paying for BBB accreditation

July 16, 2025 We have made the conscious choice not to continue paying for accreditation from the Better Business Bureau (BBB). We realize that this may raise questions among our customers, and we want to explain why we made this decision. For years, people have been told to look for BBB accredited businesses, and that it somehow reflects whether a business is on the up and up.  What most don’t realize is that businesses PAY to be accredited with the BBB. You do not EARN an accreditation- you buy it. A few months ago, an extremely negative complaint and review suddenly appeared under our name registered with the BBB.  It was a person who was upset that a Sting concert was cancelled due to fire. Their complaint was with a Music company that happened to have Cherry Tree in their name, but our business was tagged and reflected poor business practices.  We contacted the BBB many times to ask to please remove this complaint, that was obviously NOT for CherryTree Computers, from our business page.  No one at the BBB was willing or able to assist with our request…because they really don’t have the control or ability to do anything in the event of incorrect information.  This led us to then wonder…. Well, what exactly DOES the BBB do? Why would we continue to pay for accreditation if it only means we get to have the BBB logo on our website, but they don’t actually have the ability to prove or disprove how reputable a company is when it comes to business practices?  We expressed to the BBB multiple times that if the situation wasn’t rectified, we would stop paying for accreditation and let our customers know why. After a lot of waiting and no action at all from the BBB, we officially ended our relationship and will no longer pay for BBB accreditation.  We hope our services and happy customers reflect what type of business we are….and that we don’t need any special logo or stickers to prove it. Read More

We have made the decision to not continue paying for BBB accreditation Read More »

AMD’s upcoming RDNA 5 flagship could target RTX 5080-level performance with better RT

Serving tech enthusiasts for over 25 years. TechSpot means tech analysis and advice you can trust. Rumor mill: This year’s Radeon 9000 series graphics cards delivered impressive performance gains from AMD in the mid-range and mainstream market segments. However, the company chose not to compete at the very high-end categories for this generation. Although Team Red is unlikely to challenge Nvidia’s flagship products in the near future, a new GPU expected to launch next year may outperform the RTX 5080. AMD is expected to introduce a new enthusiast-class graphics card in the second half of 2026. Based on the company’s upcoming UDNA architecture, also known as RDNA 5, its configuration will closely resemble that of the Radeon RX 7900 XTX. Prominent leaker KeplerL2, who has a solid track record, speculated about the GPU’s specifications in a series of recent posts on the AnandTech forums. While the RX 9070 XT, the fastest GPU in the RDNA 4 generation, can outperform Nvidia’s GeForce RTX 5070 Ti in certain scenarios, AMD did not attempt to rival the RTX 5080, let alone the RTX 5090. However, the next lineup is expected to resemble RDNA 3 featuring a halo product that outperforms Nvidia’s 5080. The GPU won’t compete with the hypothetical RTX 6090 but could trade blows with a 6080. Similar to the 7900 XTX, the upcoming high-end AMD GPU will likely include 96 compute units and a 384-bit memory bus. A mid-range version is expected to offer 64 compute units and a 256-bit memory bus, resembling the 9070 XT. A mainstream option might be similar to the 9060 XT, with 32 compute units and a 128-bit bus. According to sources familiar with AMD’s hardware roadmap, Kepler previously estimated that UDNA will improve raster performance by approximately 20 percent over RDNA 4 and double its ray tracing capabilities. RDNA 4 already represents a significant leap in ray tracing over its predecessor. Also check out: AMD Stagnation :: Radeon 9060 XT 8GB vs 7600 vs 7600 vs 5600 XT Benchmark Our benchmarks show that the Radeon RX 9070 XT outperforms the 7900 XTX in ray tracing despite sitting an entire weight class below it in traditional rasterization. A UDNA-based GPU with the same configuration as the 7900 XTX could become a ray tracing powerhouse and may even address Radeon’s lingering disadvantage against GeForce in path tracing. Meanwhile, AMD’s UDNA architecture is also expected to power the PlayStation 6 and the next Xbox console. A recently leaked die shot suggests that Microsoft’s upcoming console includes 80 compute units, potentially outperforming the RTX 5080. With a projected price exceeding $1,000 (unlikely but that’s the rumor these days), the console appears to target the pre-built PC market instead of the traditional console market. Read More

AMD’s upcoming RDNA 5 flagship could target RTX 5080-level performance with better RT Read More »

WhatsApp is dropping its native Windows app in favor of a web-based version

Serving tech enthusiasts for over 25 years. TechSpot means tech analysis and advice you can trust. Editor’s take: Meta is preparing to deliver a worse WhatsApp experience on Windows 11 by discontinuing investment in its native desktop app. While there’s no official confirmation of this move yet, the latest WhatsApp beta makes the situation clear. The latest WhatsApp Beta introduces an unexpected change for Windows users. The update reportedly discontinues the native UWP app, replacing it with an empty shell built around the Chromium-based Edge browser framework found in recent Windows versions. WhatsApp launched a native Windows version in 2016, later converting it to use the Universal Windows Platform API with the WinUI framework. This native approach gave the app a performance edge over the web-based version. Now, Meta is returning to WebView2, the Edge framework that wraps apps around the Windows native browser component. The latest WhatsApp beta essentially behaves like the web.whatsapp.com service, which users access by pairing the mobile app with a desktop browser. By wrapping a bit of web code around the WebView2 component, WhatsApp will consume more RAM and deliver reduced performance compared to previous versions. Recent tests by Windows Latest show the new beta is consuming around 30 percent more RAM than the existing native (UWP/WebUI) stable version. Like the user-facing Edge browser, Chrome, and other Chromium-based browsers, WebView2 is a native Windows component built on the Chromium layout engine. Many simple Windows apps built around HTML, CSS, JavaScript, and other non-native web technologies rely on this component. Meta’s decision to turn back the clock with an inferior messaging experience for billions of PC users may come down to money. Windows Latest speculates that a tech giant pulling in $164.5 billion a year doesn’t want to spend a fraction of its vast wealth maintaining two separate codebases for the same app. Forcing users into a single UI benefits the company, while end users endure a worse experience on PC. Even Meta’s documentation says a native WhatsApp app offers better performance, higher reliability, and additional teamworking features – so either the developers neglected to update the docs or they simply don’t care how users feel about the UI. Another possible explanation for this potential WhatsApp fiasco is that Meta’s developers are being lazy on some desktop systems, while focusing more on the phone apps, which is exactly what they did with Facebook Messenger. The company has also drug its feet on other platforms. The company released a native iPad version just last month – a mere 15 years after Apple launched its tablet line. This patchy approach leaves PC users stuck with a downgraded experience, raising questions about Meta’s commitment to its desktop audience. Read More

WhatsApp is dropping its native Windows app in favor of a web-based version Read More »

Ryzen Threadripper Pro 9995WX scores 86% higher than its predecessor

Serving tech enthusiasts for over 25 years. TechSpot means tech analysis and advice you can trust. In a nutshell: AMD’s new Ryzen Threadripper Pro 9995WX has scored 186,800 points in the Cinebench R23 multi-core benchmark – about 86 percent higher than the 100,291 points posted by its predecessor, the Threadripper Pro 7995WX. However, the 7995WX still holds the top spot in HWBOT rankings with 210,702 points. A submission by SkywalkerAMD shows the 96-core CPU was overclocked to nearly 5 GHz on all cores during the test run, with an effective core clock of 4997.63 MHz. The chip drew a massive 947W of power while being cooled by a liquid AIO cooler. The overclocking was done manually, without using AMD’s Precision Boost Overdrive (PBO) feature. The test system included an Asus Pro WS RTX50-SAGE WIFI motherboard running BIOS version 1106. The build also featured 144GB of DDR5-6000 CL-32 G.Skill RAM. It ran Windows 11 with the 24H2 update. According to a post on Chinese forum Chiphell, the 9995WX scored 173,452 points in an earlier benchmark run. The test was conducted with PBO enabled, with the chip drawing up to 840W of power. However, the post did not mention what type of cooling was used for the test. Despite the impressive showing by the 9995WX, the highest HWBOT score still belongs to the Threadripper Pro 7995WX, which hit an astronomical 210,702 points during a 2023 test run. Liquid nitrogen cooling helped the chip reach an overclocked frequency of 6.25 GHz. How far overclockers can push the 9995WX remains unknown, but they are likely to attempt breaking the existing record soon. Online speculation suggests the new chip could hit a whopping 250,000 points on Cinebench R23, potentially claiming top honors on the HWBOT leaderboard. AMD announced five Threadripper Pro 9000WX CPUs at Computex in May, revealing pricing and availability details last week. The top SKU in the lineup, the Ryzen Threadripper Pro 9995WX, features 96 Zen 5 cores, 192 threads, a 5.4 GHz max boost clock, 384MB of L3 cache, and a 350W TDP. Team Red set the price at $11,699. The most “affordable” model, the Threadripper Pro 9955WX, features 16 cores, 32 threads, up to 5.4 GHz boost frequency, 64MB of L3 cache, and a 350W TDP. Team Red priced it at a hefty $1,649. The new CPUs will be available for DIY builders and in pre-built workstations starting July 23. Read More

Ryzen Threadripper Pro 9995WX scores 86% higher than its predecessor Read More »

AI-generated legal filings are making a mess of the judicial system

Serving tech enthusiasts for over 25 years. TechSpot means tech analysis and advice you can trust. In context: Large language models have already been used to cheat in school and spread misinformation in news reports. Now they’re creeping into the courts, fueling bogus filings that judges face amid heavy caseloads – raising new risks for a legal system already stretched thin. A recent Ars Technica report detailed a Georgia appeals court decision highlighting a growing risk for the US legal system: AI-generated hallucinations creeping into court filings and even influencing judicial rulings. In the divorce dispute, the husband’s lawyer submitted a draft order peppered with citations to cases that do not exist – likely invented by generative AI tools like ChatGPT. The initial trial court signed off on the document and subsequently ruled in the husband’s favor. Only when the wife appealed did the fabricated citations come to light. The appellate panel, led by Judge Jeff Watkins, vacated the order, noting that the bogus cases had undermined the court’s ability to review the decision. Watkins didn’t mince words, calling the citations possible generative-artificial intelligence hallucinations. The court fined the husband’s lawyer $2,500. That might sound like a one-off, but a lawyer was fined $15,000 in February under similar circumstances. Legal experts warn it is likely a sign of things to come. Generative AI tools are notoriously prone to fabricating information with convincing confidence – a behavior labeled “hallucination.” As AI becomes more accessible to both overwhelmed lawyers and self-represented litigants, experts say judges will increasingly face filings filled with fake cases, phantom precedents, and garbled legal reasoning dressed up to look legitimate. The problem is compounded by a legal system already stretched thin. In many jurisdictions, judges routinely rubberstamp orders drafted by attorneys. However, the use of AI raises the stakes. Appellate Court Opinion on False Legal Citations via Ars Technica “I can envision such a scenario in any number of situations where a trial judge maintains a heavy docket,” said John Browning, a former Texas appellate judge and legal scholar who has written extensively on AI ethics in law. Browning told Ars Technica he thinks it’s “frighteningly likely” these kinds of mistakes will become more common. He and other experts warn that courts, especially at the lower levels, are ill-prepared to handle this influx of AI-driven nonsense. Only two states – Michigan and West Virginia – currently require judges to maintain a basic level of “tech competence” when it comes to AI. Some judges have banned AI-generated filings altogether or mandated disclosure of AI use, but these policies are patchy, inconsistent, and hard to enforce due to case volume. Meanwhile, AI-generated filings aren’t always obvious. Large language models often invent realistic-sounding case names, plausible citations, and official-sounding legal jargon. Browning notes that judges can watch for telltale signs: incorrect court reporters, placeholder case numbers like “123456,” or stilted, formulaic language. However, as AI tools become more sophisticated, these giveaways may fade. Researchers, like Peter Henderson at Princeton’s Polaris Lab, are developing tools to track AI’s influence on court filings and are advocating for open repositories of legitimate case law to simplify verification. Others have floated novel solutions, such as “bounty systems” to reward those who catch fabricated cases before they slip through. For now, the Georgia divorce case stands as a cautionary tale – not just about careless lawyers, but about a court system that may be too overwhelmed to track AI use in every legal document. As Judge Watkins warned, if AI-generated hallucinations continue slipping into court records unchecked, they threaten to erode confidence in the justice system itself. Image credit: Shutterstock Read More

AI-generated legal filings are making a mess of the judicial system Read More »

VirtualBox is a free and powerful tool for running multiple operating systems

VirtualBox is a powerful x86 and AMD64/Intel64 virtualization product for enterprise as well as home use. Not only is VirtualBox an extremely feature rich, high performance product for enterprise customers, it is also the only professional solution that is freely available as Open Source Software under the terms of the GNU General Public License (GPL) version 2. Note: It has been reported that version 7.0.20 has better performance that version 7.1, so we have kept this version available for users. Version 7.1 is also listed for those interested. Can I run macOS on a Windows machine? Yes, with VirtualBox, you can install multiple operating systems on a single PC and seamlessly switch between them, including macOS on Intel hardware. VirtualBox can run multiple x86 OS such as Windows, macOS, Linux distributions, FreeBSD, and OpenBSD on your host machine. The operating systems run within an application, which virtualizes the hardware in a completely isolated environment. Is VirtualBox free? Yes, VirtualBox is a free and open source virtual machine platform for personal, educational, or evaluation use. Do I need to dual boot or repartition the disk? No, that’s not necessary. VirtualBox uses your computer’s file system and creates files that map to a virtual machine’s disk drives, so there is no need to create a partition for each operating system. If you already have another OS with dual boot, you can use VirtualBox to run the other operating system in a virtual machine on your host operating system. Instead of dual booting, you can run both operating systems simultaneously and seamlessly switch from one operating system to another with a click of your mouse. Can I run an x86 virtual machine on Arm hardware? Unfortunately, no. You can’t run an x86 image on Arm via VirtualBox. Virtual Box will only allow you to run virtual machines on the same underlying architecture as your host machine supports. Features Modularity VirtualBox has an extremely modular design with well-defined internal programming interfaces and a client/server design. This makes it easy to control it from several interfaces at once: for example, you can start a virtual machine in a typical virtual machine GUI and then control that machine from the command line, or possibly remotely. VirtualBox also comes with a full Software Development Kit: even though it is Open Source Software, you don’t have to hack the source to write a new interface for VirtualBox. Virtual machine descriptions in XML The configuration settings of virtual machines are stored entirely in XML and are independent of the local machines. Virtual machine definitions can therefore easily be ported to other computers. Guest Additions for Windows, Linux and Solaris VirtualBox has special software that can be installed inside Windows, Linux and Solaris virtual machines to improve performance and make integration much more seamless. Among the features provided by these Guest Additions are mouse pointer integration and arbitrary screen solutions (e.g. by resizing the guest window). There are also guest additions for OS/2 with somewhat reduced functionality. Shared folders Like many other virtualization solutions, for easy data exchange between hosts and guests, VirtualBox allows for declaring certain host directories as “shared folders”, which can then be accessed from within virtual machines. VirtualBox is being actively developed with frequent releases and has an ever growing list of features, supported guest operating systems and platforms it runs on. VirtualBox is a community effort backed by a dedicated company: everyone is encouraged to contribute while Oracle ensures the product always meets professional quality criteria. What’s New This is a maintenance release. The following items were fixed or added: VMM: Fixed issue when running a nested VM caused Guru Meditation for outer VM NAT: Fixed issue when VMs with long names were unable to start (github:GH-16) Linux host: Fixed possible kernel panic when using bridged networking with a network interface handled by the ixgbe driver on newer kernels Windows Host: Fixed issue resulting in BSOD upon closing VirtualBox GUI after host package uninstall (github:GH-38) Windows Host: General improvements in drivers installation Windows Host: Implement support for exposing AVX/AVX2 to the guest when Hyper-V is used (github:GH-36) Recording: Fixed issue when Windows Guest Machine was unable to start when recording was enabled in Display Settings (bug #22363) Linux Host and Guest: Added additional fixes to support kernel 6.16 Linux Guest Additions: Fixed issue when ‘rcvboxadd status-kernel’ was reporting incorrect status when guest was running kernel 3.10 series and older Linux Guest Additions: Fixed issue when VBoxClient was unable to start if guest was running kernel 2.6 series and older Linux Guest Additions: Fixed issue which caused a warning in system log due to incorrect udev rule Read More

VirtualBox is a free and powerful tool for running multiple operating systems Read More »

Scroll to Top