Relations (1)
cross_type 4.00 — strongly supporting 15 facts
GPT-4 is a proprietary large language model developed by OpenAI, as evidenced by its technical reports [1] and its classification as an OpenAI product {fact:1, fact:2, fact:6}. OpenAI continues to release and maintain iterations of this model, such as GPT-4o and GPT-4o mini {fact:5, fact:7, fact:15}.
Facts (15)
Sources
Medical Hallucination in Foundation Models and Their ... medrxiv.org 4 facts
claimOpenAI's GPT-4o model, released in May 2024, is a multimodal model capable of processing and generating text, images, and audio with enhanced reasoning and factual accuracy.
procedureThe procedure for refining text for completeness and structure involves prompting OpenAI’s GPT-4o to use text extracted by pdfminer to restore missing text from Marker-extracted content, while ensuring the final output is ordered and in Markdown format.
procedureThe procedure for providing summaries of extracted images involves using the multimodal capability of OpenAI’s GPT-4o to generate concise summaries for critical visual content in case records.
procedureThe procedure for handling missing tables in medical case records involves: (1) prompting OpenAI’s GPT-4o model to identify missing tables in text extracted by Marker, (2) re-parsing the document with Marker if the model detects missing tables, and (3) limiting this verification process to a maximum of four trials.
The Synergy of Symbolic and Connectionist AI in LLM-Empowered ... arxiv.org 2 facts
Medical Hallucination in Foundation Models and Their Impact on ... medrxiv.org 2 facts
claimOpenAI's GPT-4o is a multimodal model capable of processing and generating text, images, and audio with improved factual consistency.
measurementOpenAI released GPT-4o in May 2024 and GPT-4o mini in July 2024.
The Synergy of Symbolic and Connectionist AI in LLM ... arxiv.org 1 fact
claimOpenAI’s GPT-4 is an example of a Large Language Model that demonstrates unprecedented capabilities in natural language understanding and generation, exhibiting robust performance across a range of complex tasks.
What Is Open Source Software? - IBM ibm.com 1 fact
claimOpen source LLMs promote a transparent, accessible, and community-driven approach compared to proprietary models like Google's LaMDA and OpenAI's ChatGPT-3 and GPT-4.
vectara/hallucination-leaderboard - GitHub github.com 1 fact
referenceThe Vectara hallucination leaderboard utilizes specific API access points for various large language models: Llama 4 Maverick 17B 128E Instruct FP8 and Llama 4 Scout 17B 16E Instruct are accessed via Together AI; Microsoft Phi-4 and Phi-4-Mini are accessed via Azure; Mistral Ministral 3B, Ministral 8B, Mistral Large, Mistral Medium, and Mistral Small are accessed via Mistral AI's API; Kimi-K2-Instruct-0905 is accessed via Moonshot AI API; GPT-4.1, GPT-4o, GPT-5-High, GPT-5-Mini, GPT-5-Minimal, GPT-5-Nano, o3-Pro, o4-Mini-High, and o4-Mini-Low are accessed via OpenAI API; GPT-OSS-120B, GLM-4.5-AIR-FP8 are accessed via Together AI; Qwen3-4b, Qwen3-8b, Qwen3-14b, Qwen3-32b, and Qwen3-80b-a3b-thinking are accessed via dashscope API; Snowflake-Arctic-Instruct is accessed via Replicate API; Grok-3, Grok-4-Fast-Reasoning, and Grok-4-Fast-Non-Reasoning are accessed via xAI's API; and GLM-4.6 is accessed via deepinfra.
Real-Time Evaluation Models for RAG: Who Detects Hallucinations ... cleanlab.ai 1 fact
referenceThe FinQA dataset consists of complex questions from financial experts regarding public financial reports, with responses generated by OpenAI’s GPT-4o LLM.
Re-evaluating Hallucination Detection in LLMs - arXiv arxiv.org 1 fact
referenceThe 'GPT-4 Technical Report' by OpenAI et al. (2024) provides technical documentation and performance details for the GPT-4 large language model, published as an arXiv preprint.
Leveraging Knowledge Graphs and LLM Reasoning to Identify ... arxiv.org 1 fact
referenceThe experimental evaluation of the LLM agent framework utilized OpenAI’s GPT-4o via Langchain QA chains, interacting with a Neo4j knowledge graph through LLM-generated Cypher queries, with configuration settings of temperature 0.0, top_p 0.95, and a 4096-token limit.
LLM Hallucination Detection and Mitigation: State of the Art in 2026 zylos.ai 1 fact
claimOpenAI's 2026 research on reasoning models demonstrates that naturally understandable chain-of-thought reasoning traces are reinforced through reinforcement learning, and that simple prompted GPT-4o models can effectively monitor for reward hacking in frontier reasoning models like o1 and o3-mini successors.