Article
Version 1
Preserved in Portico This version is not peer-reviewed
Masterminding the Giant : Shannonian Entropic Threshold Perplexity AI
Version 1
: Received: 19 May 2024 / Approved: 20 May 2024 / Online: 20 May 2024 (11:03:46 CEST)
How to cite: A Mageed, D. I. Masterminding the Giant : Shannonian Entropic Threshold Perplexity AI. Preprints 2024, 2024051274. https://doi.org/10.20944/preprints202405.1274.v1 A Mageed, D. I. Masterminding the Giant : Shannonian Entropic Threshold Perplexity AI. Preprints 2024, 2024051274. https://doi.org/10.20944/preprints202405.1274.v1
Abstract
Undoubtedly, there is a potential impact of large language models like ChatGPT in revolutionizing information retrieval and knowledge discovery, particularly in the context of the vast amount of electronic material available. On another strong note, the current work offers a first time ever mathematical approach to accurately fine tune perplexity AI, which sets a giant step ahead towards accurate next generations AI. Having started this world leading discovery, the upper and lower bounds of perplexity AI, namely , , respectively, are obtained for the first-time ever. Concluding remarks combined with emerging open problems and next research phase are provided.
Keywords
Perplexity, AI, ChatGPT, Entropy, Natural Language Processing(NLP), Increasing/Decreasing test, or (IDT), Threshold.
Subject
Computer Science and Mathematics, Artificial Intelligence and Machine Learning
Copyright: This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Comments (0)
We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.
Leave a public commentSend a private comment to the author(s)
* All users must log in before leaving a comment