In a database, you have entities (which have attributes), and relationships between those entities. Managing them is key to preventing chaos from engulfing your database, which is where the concept of keys comes in. These unique identifiers enable you to pick specific rows in an entity set, as well as define their relationships to rows in other entity sets, allowing your database to handle complex computations.
Let’s explore keys in DBMS (database management systems) in more detail, before digging into everything you need to know about the most important keys – primary keys.
Understanding Keys in DBMS
Keys in DBMS are attributes that you use to identify specific rows inside a table, in addition to finding the relation between two tables. For example, let’s say you have a table for students, with that table recording each student’s “ID Number,” “Name,” “Address,” and “Teacher” as attributes. If you want to identify a specific student in the table, you’ll need to use one of these attributes as a key that allows you to pull the student’s record from your database. In this case “ID Number” is likely the best choice because it’s a unique attribute that only applies to a single student.
Types of Keys in DBMS
Beyond the basics of serving as unique identifiers for rows in a database, keys in DBMS can take several forms:
- Primary Keys – An attribute that is present in the table for all of the records it contains, with each instance of that attribute being unique to the record. The previously-mentioned “ID Number” for students is a great example, as no student can have the same number as another student.
- Foreign Key – Foreign keys allow you to define and establish relationships between a pair of tables. If Table A needs to refer to the primary key in Table B, you’ll use a foreign key in Table A so you have values in that table to match those in Table B.
- Unique Key – These are very similar to primary keys in that both contain unique identifiers for the records in a table. The only difference is that a unique key can contain a null value, whereas a primary key can’t.
- Candidate Key – Though you may have picked a unique attribute to serve as your primary key, there may be other candidates within a table. Coming back to the student example, you may record the phone numbers and email addresses of your students, which can be as unique as the student ID assigned to the individual. These candidate keys are also unique identifiers, allowing them to be used in tandem with a primary key to identify a specific row in a table.
- Composite Key – If you have attributes that wouldn’t be unique when taken alone, but can be combined to form a unique identifier for a record, you have a composite key.
- Super Key – This term refers to the collection of attributes that uniquely identify a record, meaning it’s a combination of candidate keys. Just like an employer sifting through job candidates to find the perfect person, you’ll sift through your super key set to choose the ideal primary key amongst your candidate keys.
So, why are keys in DBMS so important?
Keys ensure you maintain data integrity across all of the tables that make up your database. Without them, the relationships between each table become messy hodgepodges, creating the potential for duplicate records and errors that deliver inaccurate reports from the database. Having unique identifiers (in the form of keys) allows you to be certain that any record you pull, and the relationships that apply to that record, are accurate and unrepeated.
Primary Key Essentials
As mentioned, any unique attribute in a table can serve as a primary key, though this doesn’t mean that every unique attribute is a great choice. The following characteristics help you to define the perfect primary key.
Uniqueness
If your primary key is repeatable across records, it can’t serve as a unique identifier for a single record. For example, our student table may have multiple people named “John,” so you can’t use the “Name” attribute to find a specific student. You need something unique to that student, such as the previously mentioned ID number.
Non-Null Values
Primary keys must always contain a value, else you risk losing records in a table because you have no way of calling upon them. This need for non-null values can be used to eliminate some candidates from primary key content. For instance, it’s feasible (though unlikely) that a student won’t have an email address, creating the potential for null values that mean the email address attribute can’t be a primary key.
Immutability
A primary key that can change over time is a key that can cause confusion. Immutability is the term used for any attribute that’s unchanging to the point where it’s an evergreen attribute that you can use to identify a specific record forever.
Minimal
Ideally, one table should have one attribute that serves as its primary key, which is where the term “minimal” comes in. It’s possible for a table to have a composite or super key set, though both create the possibility of confusion and data integrity issues.
The Importance of a Primary Key in DBMS
We can distill the reason why having a primary key in DBMS for each of your tables is important into the following reasons:
- You can use a primary key to identify each unique record in a table, meaning no multi-result returns to your database searches.
- Having a primary key means a record can’t be repeated in the table.
- Primary keys make data retrieval more efficient because you can use a single attribute for searches rather than multiple.
Functions of Primary Keys
Primary keys in DBMS serve several functions, each of which is critical to your DBMS.
Data Identification
Imagine walking into a crowded room and shouting out a name. The odds are that several people (all of whom have the same name) will turn their heads to look at you. That’s basically what you’re doing if you try to pull records from a table without using a primary key.
A primary key in DBMS serves as a unique identifier that you can use to pull specific records. Coming back to the student example mentioned earlier, a “Student ID” is only applicable to a single student, making it a unique identifier you can use to find that student in your database.
Ensure Data Integrity
Primary keys protect data integrity in two ways.
First, they prevent duplicate records from building up inside a single table, ensuring you don’t get multiple instances of the same record. Second, they ensure referential integrity, which is the term used to describe what happens when one table in your database needs to refer to the records stored in another table.
For example, let’s say you have tables for “Students” and “Teachers” in your database. The primary keys assigned to your students and teachers allow you to pull individual records as needed from each table. But every “Teacher” has multiple “Students” in their class. So, your primary key from the “Students” table is used as a foreign key in the “Teachers” table, allowing you to denote the one-to-many relationship between a teacher and their class of students. That foreign key also ensures referential integrity because it contains the unique identifiers for students, which you can look up in your “Students” table.
Data Retrieval
If you need to pull a specific record from a table, you can’t rely on attributes that can repeat across several records in that table. Again, the “Name” example highlights the problem here, as several people could have the same name. You need a unique identifier for each record so you can retrieve a single record from a huge set without having to pore through hundreds (or even thousands) of records.
Best Practices for Primary Key Selection
Now that you understand how primary keys in DBMS work, here are some best practices for selecting the right primary key for your table:
- Choose Appropriate Attributes as Candidates – If the attribute isn’t unique to each record, or it can contain a null value (as is the case with email addresses and phone numbers), it’s not a good candidate for a primary key.
- Avoid Using Sensitive Information – Using personal or sensitive information as a primary key creates a security risk because anybody who cracks your database could use that information for other purposes. Make your primary keys unique, and only applicable, to your database, which allows you to encrypt any sensitive information stored in your tables.
- Consider Surrogate Keys – Some tables don’t have natural attributes that you can use as primary keys. In these cases, you can create a primary key out of thin air and assign it to each record. The “Student ID” referenced earlier is a great example, as students entering a school don’t come with their own ID numbers. Those numbers are given to the student (or simply used in the database that collects their data), making them surrogate keys.
- Ensure Primary Key Stability – Any attribute that can change isn’t suitable for use as a primary key because it causes stability issues. Names, email addresses, phone numbers, and even bank account details are all things that can change, making them unsuitable. Evergreen and unchanging is the way to go with primary keys.
Choose the Right Keys for Your Database
You need to understand the importance of a primary key in DBMS (or multiple primary keys when you have several tables) so you can define the relationships between tables and identify unique records inside your tables. Without primary keys, you’ll find it much harder to run reports because you won’t feel confident in the accuracy of the data returned. Each search may pull up duplicate or incorrect records because of a lack of unique identifiers.
Thankfully, many of the tables you create will have attributes that lend themselves well to primary key status. And even when that isn’t the case, you can use surrogate keys in DBMS to assign primary keys to your tables. Experiment with your databases, testing different potential primary keys to see what works best for you.
Related posts
Source:
- Agenda Digitale, published on November 25th, 2025
In recent years, the word ” sustainability ” has become a firm fixture in the corporate lexicon. However, simply “doing no harm” is no longer enough: the climate crisis , social inequalities , and the erosion of natural resources require a change of pace. This is where the net-positive paradigm comes in , a model that isn’t content to simply reduce negative impacts, but aims to generate more social and environmental value than is consumed.
This isn’t about philanthropy, nor is it about reputational makeovers: net-positive is a strategic approach that intertwines economics, technology, and corporate culture. Within this framework, digitalization becomes an essential lever, capable of enabling regenerative models through circular platforms and exponential technologies.
Blockchain, AI, and IoT: The Technological Triad of Regeneration
Blockchain, Artificial Intelligence, and the Internet of Things represent the technological triad that makes this paradigm shift possible. Each addresses a critical point in regeneration.
Blockchain guarantees the traceability of material flows and product life cycles, allowing a regenerated dress or a bottle collected at sea to tell their story in a transparent and verifiable way.
Artificial Intelligence optimizes recovery and redistribution chains, predicting supply and demand, reducing waste and improving the efficiency of circular processes .
Finally, IoT enables real-time monitoring, from sensors installed at recycling plants to sharing mobility platforms, returning granular data for quick, informed decisions.
These integrated technologies allow us to move beyond linear vision and enable systems in which value is continuously regenerated.
New business models: from product-as-a-service to incentive tokens
Digital regeneration is n’t limited to the technological dimension; it’s redefining business models. More and more companies are adopting product-as-a-service approaches , transforming goods into services: from technical clothing rentals to pay-per-use for industrial machinery. This approach reduces resource consumption and encourages modular design, designed for reuse.
At the same time, circular marketplaces create ecosystems where materials, components, and products find new life. No longer waste, but input for other production processes. The logic of scarcity is overturned in an economy of regenerated abundance.
To complete the picture, incentive tokens — digital tools that reward virtuous behavior, from collecting plastic from the sea to reusing used clothing — activate global communities and catalyze private capital for regeneration.
Measuring Impact: Integrated Metrics for Net-Positiveness
One of the main obstacles to the widespread adoption of net-positive models is the difficulty of measuring their impact. Traditional profit-focused accounting systems are not enough. They need to be combined with integrated metrics that combine ESG and ROI, such as impact-weighted accounting or innovative indicators like lifetime carbon savings.
In this way, companies can validate the scalability of their models and attract investors who are increasingly attentive to financial returns that go hand in hand with social and environmental returns.
Case studies: RePlanet Energy, RIFO, and Ogyre
Concrete examples demonstrate how the combination of circular platforms and exponential technologies can generate real value. RePlanet Energy has defined its Massive Transformative Purpose as “Enabling Regeneration” and is now providing sustainable energy to Nigerian schools and hospitals, thanks in part to transparent blockchain-based supply chains and the active contribution of employees. RIFO, a Tuscan circular fashion brand, regenerates textile waste into new clothing, supporting local artisans and promoting workplace inclusion, with transparency in the production process as a distinctive feature and driver of loyalty. Ogyre incentivizes fishermen to collect plastic during their fishing trips; the recovered material is digitally tracked and transformed into new products, while the global community participates through tokens and environmental compensation programs.
These cases demonstrate how regeneration and profitability are not contradictory, but can actually feed off each other, strengthening the competitiveness of businesses.
From Net Zero to Net Positive: The Role of Massive Transformative Purpose
The crucial point lies in the distinction between sustainability and regeneration. The former aims for net zero, that is, reducing the impact until it is completely neutralized. The latter goes further, aiming for a net positive, capable of giving back more than it consumes.
This shift in perspective requires a strong Massive Transformative Purpose: an inspiring and shared goal that guides strategic choices, preventing technology from becoming a sterile end. Without this level of intentionality, even the most advanced tools risk turning into gadgets with no impact.
Regenerating business also means regenerating skills to train a new generation of professionals capable not only of using technologies but also of directing them towards regenerative business models. From this perspective, training becomes the first step in a transformation that is simultaneously cultural, economic, and social.
The Regenerative Future: Technology, Skills, and Shared Value
Digital regeneration is not an abstract concept, but a concrete practice already being tested by companies in Europe and around the world. It’s an opportunity for businesses to redefine their role, moving from mere economic operators to drivers of net-positive value for society and the environment.
The combination of blockchain, AI, and IoT with circular product-as-a-service models, marketplaces, and incentive tokens can enable scalable and sustainable regenerative ecosystems. The future of business isn’t just measured in terms of margins, but in the ability to leave the world better than we found it.
Source:
- Raconteur, published on November 06th, 2025
Many firms have conducted successful Artificial Intelligence (AI) pilot projects, but scaling them across departments and workflows remains a challenge. Inference costs, data silos, talent gaps and poor alignment with business strategy are just some of the issues that leave organisations trapped in pilot purgatory. This inability to scale successful experiments means AI’s potential for improving enterprise efficiency, decision-making and innovation isn’t fully realised. So what’s the solution?
Although it’s not a magic bullet, an AI operating model is really the foundation for scaling pilot projects up to enterprise-wide deployments. Essentially it’s a structured framework that defines how the organisation develops, deploys and governs AI. By bringing together infrastructure, data, people, and governance in a flexible and secure way, it ensures that AI delivers value at scale while remaining ethical and compliant.
“A successful AI proof-of-concept is like building a single race car that can go fast,” says Professor Yu Xiong, chair of business analytics at the UK-based Surrey Business School. “An efficient AI technology operations model, however, is the entire system – the processes, tools, and team structures – for continuously manufacturing, maintaining, and safely operating an entire fleet of cars.”
But while the importance of this framework is clear, how should enterprises establish and embed it?
“It begins with a clear strategy that defines objectives, desired outcomes, and measurable success criteria, such as model performance, bias detection, and regulatory compliance metrics,” says Professor Azadeh Haratiannezhadi, co-founder of generative AI company Taktify and professor of generative AI in cybersecurity at OPIT – the Open Institute of Technology.
Platforms, tools and MLOps pipelines that enable models to be deployed, monitored and scaled in a safe and efficient way are also essential in practical terms.
“Tools and infrastructure must also be selected with transparency, cost, and governance in mind,” says Efrain Ruh, continental chief technology officer for Europe at Digitate. “Crucially, organisations need to continuously monitor the evolving AI landscape and adapt their models to new capabilities and market offerings.”
An open approach
The most effective AI operating models are also founded on openness, interoperability and modularity. Open source platforms and tools provide greater control over data, deployment environments and costs, for example. These characteristics can help enterprises to avoid vendor lock-in, successfully align AI to business culture and values, and embed it safely into cross-department workflows.
“Modularity and platformisation…avoids building isolated ‘silos’ for each project,” explains professor Xiong. “Instead, it provides a shared, reusable ‘AI platform’ that integrates toolchains for data preparation, model training, deployment, monitoring, and retraining. This drastically improves efficiency and reduces the cost of redundant work.”
A strong data strategy is equally vital for ensuring high-quality performance and reducing bias. Ideally, the AI operating model should be cloud and LLM agnostic too.
“This allows organisations to coordinate and orchestrate AI agents from various sources, whether that’s internal or 3rd party,” says Babak Hodjat, global chief technology officer of AI at Cognizant. “The interoperability also means businesses can adopt an agile iterative process for AI projects that is guided by measuring efficiency, productivity, and quality gains, while guaranteeing trust and safety are built into all elements of design and implementation.”
A robust AI operating model should feature clear objectives for compliance, security and data privacy, as well as accountability structures. Richard Corbridge, chief information officer of Segro, advises organisations to: “Start small with well-scoped pilots that solve real pain points, then bake in repeatable patterns, data contracts, test harnesses, explainability checks and rollback plans, so learning can be scaled without multiplying risk. If you don’t codify how models are approved, deployed, monitored and retired, you won’t get past pilot purgatory.”
Of course, technology alone can’t drive successful AI adoption at scale: the right skills and culture are also essential for embedding AI across the enterprise.
“Multidisciplinary teams that combine technical expertise in AI, security, and governance with deep business knowledge create a foundation for sustainable adoption,” says Professor Haratiannezhadi. “Ongoing training ensures staff acquire advanced AI skills while understanding associated risks and responsibilities.”
Ultimately, an AI operating model is the playbook that enables an enterprise to use AI responsibly and effectively at scale. By drawing together governance, technological infrastructure, cultural change and open collaboration, it supports the shift from isolated experiments to the kind of sustainable AI capability that can drive competitive advantage.
In other words, it’s the foundation for turning ambition into reality, and finally escaping pilot purgatory for good.
Have questions?
Visit our FAQ page or get in touch with us!
Write us at +39 335 576 0263
Get in touch at hello@opit.com
Talk to one of our Study Advisors
We are international
We can speak in: