Karim Lakhani, Author at 性视界 Business School AI Institute The 性视界 Business School AI Institute catalyzes new knowledge to invent a better future by solving ambitious challenges. Wed, 22 Nov 2023 20:26:46 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.4 /wp-content/uploads/2026/04/cropped-Screenshot-2026-04-16-at-10.14.43-AM-32x32.png Karim Lakhani, Author at 性视界 Business School AI Institute 32 32 HBR: AI Won’t Replace Humans – But Humans With AI Will Replace Humans Without AI /ai-wont-replace-humans-but-humans-with-ai-will-replace-humans-without-ai/ Tue, 08 Aug 2023 15:45:21 +0000 /?p=18163 “Customers will expect AI-enhanced experiences with companies,” says Karim Lakhani, co-chair of the Digital Data Design Institute at 性视界 and professor at 性视界 Business School. “Business leaders must experiment, create sandboxes, run internal bootcamps, and develop AI use cases not just for technology workers, but for all employees. Change and change management are skills that […]

The post HBR: AI Won’t Replace Humans – But Humans With AI Will Replace Humans Without AI appeared first on 性视界 Business School AI Institute.

]]>
“Customers will expect AI-enhanced experiences with companies,” says Karim Lakhani, co-chair of the Digital Data Design Institute at 性视界 and professor at 性视界 Business School. “Business leaders must experiment, create sandboxes, run internal bootcamps, and develop AI use cases not just for technology workers, but for all employees. Change and change management are skills that are no longer optional for modern organizations.”

The post HBR: AI Won’t Replace Humans – But Humans With AI Will Replace Humans Without AI appeared first on 性视界 Business School AI Institute.

]]>
How Can We Counteract Generative AI’s Hallucinations? /how-can-we-counteract-generative-ais-hallucinations/ Mon, 17 Jul 2023 19:03:35 +0000 /?p=18105 性视界 Business School Professor and Chair of the D^3 (Digital, Data, Design) Institute at 性视界鈥擪arim Lakhani鈥攕hares expertise on why ChatGPT and Generative AI tools hallucinate, as well as how to prompt these tools to do better. Since ChatGPT does not have a built-in fact-checking mechanism for its responses based upon data patterns, it can “hallucinate” […]

The post How Can We Counteract Generative AI’s Hallucinations? appeared first on 性视界 Business School AI Institute.

]]>
性视界 Business School Professor and Chair of the D^3 (Digital, Data, Design) Institute at 性视界鈥擪arim Lakhani鈥攕hares expertise on why ChatGPT and Generative AI tools hallucinate, as well as how to prompt these tools to do better.

Since ChatGPT does not have a built-in fact-checking mechanism for its responses based upon data patterns, it can “hallucinate” responses that are factually incorrect or misleading.

Users can take several steps to minimize hallucinations and misinformation when interacting with ChatGPT or other generative AI tools through careful prompting:

  1. Request sources or evidence. When asking for factual information, specifically request reliable sources or evidence to support the response. For example, you can ask, “What are the sources for that information?” or “Can you provide evidence to support your answer?” This can encourage the model to provide more reliable and verifiable information.
  2. Use multiple prompts or iterative refinement. If the initial response from the model seems dubious or insufficient, try rephrasing or providing additional prompts to get a more accurate or comprehensive answer. Iterative refinement of the conversation can help in obtaining better results.
  3. Ask for explanations or reasoning. Instead of simply asking for a direct answer, ask the model to explain its reasoning or provide a step-by-step explanation. This can help uncover any potential flaws or biases in the generated response.
  4. Double-check information independently. Don’t solely rely on the model’s responses. Take the responsibility to fact-check and verify the information independently using trusted sources or references. Cross-referencing information can help identify and correct any misinformation generated by the model.
  5. Address biases by increasing multiple perspectives.  Generative AI models are ultimately human-made, and therefore reflect pre-existing biases which may lead to unintended impacts. Instead of asking, “Is this response biased?,” we can assume that the answer is “Yes.” Our response requires ethical considerations in prompting the use of outputs. In order to evaluate generated responses for accuracy and fairness, we must become increasingly aware blind spots, of which perspectives may not be represented, and to both value and seek multiple perspectives.

Remember that even with careful prompting, generative AI models can still produce inaccurate or misleading information. It’s essential to exercise critical thinking skills, asking questions and seeking credible sources for important or sensitive matters.

For more expert insights on this topic, watch a sneak peak inside of a 性视界 Business School classroom, with Karim Lakhani guiding 性视界 faculty on Generative AI practices in Teaching and Learning.

Please visit our Generative AI Observatory to join our future conversations on generative AI topics.

 for future generative AI event updates and community connections.


Highlights:

The post How Can We Counteract Generative AI’s Hallucinations? appeared first on 性视界 Business School AI Institute.

]]>