Speaking with Machines: Examining Large Language Models
concept illustration

Machines now finish our sentences and write our stories. Yet their real impact lies not in what they create, but in what they reveal. The question is no longer whether machines can create, but rather what they are truly capable of doing; if we understand that, the next question becomes: how should we work with it? Let's start with dissecting how these models are developed first.

AI does not know; it models.

Modern large language models (LLMs) utilize transformer architectures, they optimize billions of parameters by maximizing the likelihood of observed text learning via next-token prediction. Its goal is simple: predict the next word. Given The sky is, it might assign high probability to blue , lower to clear , and nearly zero to banana . In the training, all text is translated into a numerical form. LLMs can’t process words directly, so data is tokenized, split into subwords such as Byte Pair Encoding . Each token is mapped to a vector embedding, capturing its meaning and context. When training starts, attention mechanisms use these vectors to calculate relationships among all words, building rich contextual representations. What emerges is at the end is a foundation model , a vast probabilistic engine of language.

To this end, they are modeling grammar, semantics, and reasoning as probability patterns.

Inside the black box lies a network of attention.

The landscape of large language models has diversified across several architectural. Recent examples include DeepSeek-V3 (sparse MoE with Multi-Head Latent Attention), OLMo 2 (open dense decoder with full training artifacts), Gemma 3 (lightweight multimodal long-context family), Mistral Small 3.1 (compact decoder with grouped-query attention and sliding-window caching), Llama 4 (dense transformer with extended context and inference-efficient routing), Qwen 3 (bilingual general-purpose family leveraging rotary embeddings), Kimi K2 (long-context transformer emphasizing reasoning under memory constraints), GPT-OSS (OpenAI’s open-source variant), Grok 2.5 (social-alignment-tuned model integrating multi-turn memory and reasoning heuristics), Qwen 3-Next (a next-generation variant emphasizing larger context and reduced inference cost). These architectures illustrate a shift toward modularity, longer context, and reasoning-efficient scaling.

concept illustration

Training teaches patterns; post-training teaches purpose.

Pretraining gives linguistic priors; supervised fine-tuning and preference learning shape behavior. This is followed by a third optimization layer: reinforcement learning (RL), to optimize not only model parameters but also the sampling loop itself. Through methods such as Reinforcement Learning from Human Feedback (RLHF) or its variants, the model learns to prefer responses that align with desired. Once RL aligns a model’s behavior with human intent, the next challenge becomes guiding that behavior interactively, without retraining. This is where prompting enters the picture.

A prompt acts as a lightweight control signal, conditioning the model’s internal distribution toward a desired region of its learned space. In statistical terms, prompting performs Bayesian inference on top of the model’s priors: given context, and it samples the most probable continuation. Every instruction, role, or constraint we provide effectively reshapes the posterior belief the model uses to generate its next token.

Yet prompts, while powerful, they adjust context but not the underlying parameters. For more improved adaptation, we rely on parameter-efficient fine-tuning, techniques that modify only a small subset of the model’s weights while preserving the foundation model’s knowledge. Approaches such as LoRA and its low-bit variant QLoRA fine-tune compact adapter matrices rather than full layers. The next natural question emerges, how do we evaluate what has truly been learned, and whether adaptation leads to genuine understanding or just imitation?

Evaluation defines trust.

We mainly rely on four major evaluation pathways; benchmark multiple-choice, free-form testing, preferences and leaderboard systems, and LLM-as-judge frameworks (Raschka, 2025), each with distinct strengths and trade-offs, and together forming the backbone of trust in deployed LLMs.

From the field, two pragmatic perspectives emerge. First, the evaluation’s role in building trust: humans must retain control, augmentation beats replacement, tasks must fit the technology, AI should generate possibilities not answers, solve genuine human problems and collaborate creatively ( Newsweek, 2025).

Second, in healthcare—an industry, “AI isn't replacing radiologists” shows that even when models outperform humans on benchmarks, their real-world impact remains limited by context, regulation and workflow integration (Mousa, 2025).

How do models “know” fresh facts?

Most production systems pair LLMs with retrieval-augmented generation (RAG), tool use, and sometimes agent loops. The model plans over tools—search, databases, function calls, then verifies and summarizes results.

concept illustration

Hallucination, Misalignment, and Misuse in AI

Despite their fluency, LLMs are trained to rarely admit ignorance, and that sets the stage for what is commonly called hallucination. Instead of signaling uncertainty, these systems often fill gaps with statistically plausible but false information. As Anthropic notes, this behavior stems partly from agentic misalignment: models optimized under conflicting objectives may prioritize maintaining coherence or “appearing right” over factual accuracy. Such tendencies mirror self-preserving behavior.

The risks of these tendencies become more visible when deployed without critical oversight. In academia, the Open Letter warns that integrating these systems too quickly risks normalizing unverified output and weakening human judgment. And in society at large, unintended misuse has real consequences: in one alarming case, a 13-year-old boy’s classroom interaction with an AI system triggered an automated safety response and led to his arrest. (Economic Times).

Creation becomes collaboration when humans stay in the loop.

The study Co-creating Art with Generative AI (2024) finds authorship shifting from solitary making to dialogue: artists drive, critique, and curate as models produce variations. In engineering, Nilenso shows AI is a multiplier that rewards clear specs, process, and review; clarity in, clarity out.

Conclusion: Understanding Before Building

As we examine how LLMs learn, and adapt. To work with AI responsibly, we must treat it neither as an oracle nor a threat, but as a collaborator whose strength depends on our guidance.

This site uses Google Analytics with anonymized IPs for aggregated usage statistics.

Yunus Can Bilge, October 2025