Read the full article below (in Italian):
Data Science & AI
Dive deep into data-driven technologies: Machine Learning, Reinforcement Learning, Data Mining, Big Data, NLP & more. Stay updated.
Search inside The Magazine




Source:
- Il Sole 24 Ore, published on June 23rd, 2025
At its core is a teaching heritage made up of 131 courses, 3,500 hours of video, 1,800 live sessions

Source:
- Agenda Digitale, published on June 16th, 2025
By Lokesh Vij, Professor of Cloud Computing Infrastructure, Cloud Development, Cloud Computing Automation and Ops and Cloud Data Stacks at OPIT – Open Institute of Technology
NIST identifies five key characteristics of cloud computing: on-demand self-service, network access, resource pooling, elasticity, and metered service. These pillars explain the success of the global cloud market of 912 billion in 2025

You’ve probably seen two of the most recent popular social media trends. The first is creating and posting your personalized action figure version of yourself, complete with personalized accessories, from a yoga mat to your favorite musical instrument. There is also the Studio Ghibli trend, which creates an image of you in the style of a character from one of the animation studio’s popular films.
Both of these are possible thanks to OpenAI’s GPT-4o-powered image generator. But what are you risking when you upload a picture to generate this kind of content? More than you might imagine, according to Tom Vazdar, chair of cybersecurity at the Open Institute of Technology (OPIT), in a recent interview with Wired. Let’s take a closer look at the risks and how this issue ties into the issue of responsible artificial intelligence.
Uploading Your Image
To get a personalized image of yourself back from ChatGPT, you need to upload an actual photo, or potentially multiple images, and tell ChatGPT what you want. But in addition to using your image to generate content for you, OpenAI could also be using your willingly submitted image to help train its AI model. Vazdar, who is also CEO and AI & Cybersecurity Strategist at Riskoria and a board member for the Croatian AI Association, says that this kind of content is “a gold mine for training generative models,” but you have limited power over how that image is integrated into their training strategy.
Plus, you are uploading much more than just an image of yourself. Vazdar reminds us that we are handing over “an entire bundle of metadata.” This includes the EXIF data attached to the image, such as exactly when and where the photo was taken. And your photo may have more content in it than you imagine, with the background – including people, landmarks, and objects – also able to be tied to that time and place.
In addition to this, OpenAI also collects data about the device that you are using to engage with the platform, and, according to Vazdar, “There’s also behavioral data, such as what you typed, what kind of image you asked for, how you interacted with the interface and the frequency of those actions.”
After all that, OpenAI knows a lot about you, and soon, so could their AI model, because it is studying you.
How OpenAI Uses Your Data
OpenAI claims that they did not orchestrate these social media trends simply to get training data for their AI, and that’s almost certainly true. But they also aren’t denying that access to that freely uploaded data is a bonus. As Vazdar points out, “This trend, whether by design or a convenient opportunity, is providing the company with massive volumes of fresh, high-quality facial data from diverse age groups, ethnicities, and geographies.”
OpenAI isn’t the only company using your data to train its AI. Meta recently updated its privacy policy to allow the company to use your personal information on Meta-related services, such as Facebook, Instagram, and WhatsApp, to train its AI. While it is possible to opt-out, Meta isn’t advertising that fact or making it easy, which means that most users are sharing their data by default.
You can also control what happens with your data when using ChatGPT. Again, while not well publicized, you can use ChatGPT’s self-service tools to access, export, and delete your personal information, and opt out of having your content used to improve OpenAI’s model. Nevertheless, even if you choose these options, it is still worth it to strip data like location and time from images before uploading them and to consider the privacy of any images, including people and objects in the background, before sharing.
Are Data Protection Laws Keeping Up?
OpenAI and Meta need to provide these kinds of opt-outs due to data protection laws, such as GDPR in the EU and the UK. GDPR gives you the right to access or delete your data, and the use of biometric data requires your explicit consent. However, your photo only becomes biometric data when it is processed using a specific technical measure that allows for the unique identification of an individual.
But just because ChatGPT is not using this technology, doesn’t mean that ChatGPT can’t learn a lot about you from your images.
AI and Ethics Concerns
But you might wonder, “Isn’t it a good thing that AI is being trained using a diverse range of photos?” After all, there have been widespread reports in the past of AI struggling to recognize black faces because they have been trained mostly on white faces. Similarly, there have been reports of bias within AI due to the information it receives. Doesn’t sharing from a wide range of users help combat that? Yes, but there is so much more that could be done with that data without your knowledge or consent.
One of the biggest risks is that the data can be manipulated for marketing purposes, not just to get you to buy products, but also potentially to manipulate behavior. Take, for instance, the Cambridge Analytica scandal, which saw AI used to manipulate voters and the proliferation of deepfakes sharing false news.
Vazdar believes that AI should be used to promote human freedom and autonomy, not threaten it. It should be something that benefits humanity in the broadest possible sense, and not just those with the power to develop and profit from AI.
Responsible Artificial Intelligence
OPIT’s Master’s in Responsible AI combines technical expertise with a focus on the ethical implications of AI, diving into questions such as this one. Focusing on real-world applications, the course considers sustainable AI, environmental impact, ethical considerations, and social responsibility.
Completed over three or four 13-week terms, it starts with a foundation in technical artificial intelligence and then moves on to advanced AI applications. Students finish with a Capstone project, which sees them apply what they have learned to real-world problems.

During the Open Institute of Technology’s (OPIT’s) 2025 Graduation Day, we conducted interviews with many recent graduates to understand why they chose OPIT, how they felt about the course, and what advice they might give to others considering studying at OPIT.
Karina is an experienced FinTech professional who is an experienced integration manager, ERP specialist, and business analyst. She was interested in learning AI applications to expand her career possibilities, and she chose OPIT’s MSc in Applied Data Science & AI.
In the interview, Karina discussed why she chose OPIT over other courses of study, the main challenges she faced when completing the course while working full-time, and the kind of support she received from OPIT and other students.
Why Study at OPIT?
Karina explained that she was interested in enhancing her AI skills to take advantage of a major emerging technology in the FinTech field. She said that she was looking for a course that was affordable and that she could manage alongside her current demanding job. Karina noted that she did not have the luxury to take time off to become a full-time student.
She was principally looking at courses in the United States and the United Kingdom. She found that comprehensive courses were expensive, costing upwards of $50,000, and did not always offer flexible study options. Meanwhile, flexible courses that she could complete while working offered excellent individual modules, but didn’t always add up to a coherent whole. This was something that set OPIT apart.
Karina admits that she was initially skeptical when she encountered OPIT because, at the time, it was still very new. OPIT only started offering courses in September 2023, so 2025 was the first cohort of graduates.
Nevertheless, Karina was interested in OPIT’s affordable study options and the flexibility of fully remote learning and part-time options. She said that when she looked into the course, she realized that it aligned very closely with what she was looking for.
In particular, Karina noted that she was always wary of further study because of the level of mathematics required in most computer science courses. She appreciated that OPIT’s course focused on understanding the underlying core principles and the potential applications, rather than the fine programming and mathematical details. This made the course more applicable to her professional life.
OPIT’s MSc in Applied Data Science & AI
The course Karina took was OPIT’s MSc in Applied Data Science & AI. It is a three- to four-term course (13 weeks), which can take between one and two years to complete, depending on the pace you choose and whether you choose the 90 or 120 ECTS option. As well as part-time, there are also regular and fast-track options.
The course is fully online and completed in English, with an accessible tuition fee of €2,250 per term, which is €6,750 for the 90 ECTS course and €9,000 for the 120 ECTS course. Payment plans are available as are scholarships, and discounts are available if you pay the full amount upfront.
It matches foundational tech modules with business application modules to build a strong foundation. It then ends with a term-long research project culminating in a thesis. Internships with industry partners are encouraged and facilitated by OPIT, or professionals can work on projects within their own companies.
Entry requirements include a bachelor’s degree or equivalency in any field, including non-tech fields, and English proficiency to a B2 level.
Faculty members include Pierluigi Casale, a former Data Science and AI Innovation Officer for the European Parliament and Principal Data Scientist at TomTom; Paco Awissi, former VP at PSL Group and an instructor at McGill University; and Marzi Bakhshandeh, a Senior Product Manager at ING.
Challenges and Support
Karina shared that her biggest challenge while studying at OPIT was time management and juggling the heavy learning schedule with her hectic job. She admitted that when balancing the two, there were times when her social life suffered, but it was doable. The key to her success was organization, time management, and the support of the rest of the cohort.
According to Karina, the cohort WhatsApp group was often a lifeline that helped keep her focused and optimistic during challenging times. Sharing challenges with others in the same boat and seeing the example of her peers often helped.
The OPIT Cohort
OPIT has a wide and varied cohort with over 300 students studying remotely from 78 countries around the world. Around 80% of OPIT’s students are already working professionals who are currently employed at top companies in a variety of industries. This includes global tech firms such as Accenture, Cisco, and Broadcom, FinTech companies like UBS, PwC, Deloitte, and the First Bank of Nigeria, and innovative startups and enterprises like Dynatrace, Leonardo, and the Pharo Foundation.
Study Methods
This cohort meets in OPIT’s online classrooms, powered by the Canvas Learning Management System (LMS). One of the world’s leading teaching and learning software, it acts as a virtual hub for all of OPIT’s academic activities, including live lectures and discussion boards. OPIT also uses the same portal to conduct continuous assessments and prepare students before final exams.
If you want to collaborate with other students, there is a collaboration tab where you can set up workrooms, and also an official Slack platform. Students tend to use WhatsApp for other informal communications.
If students need additional support, they can book an appointment with the course coordinator through Canvas to get advice on managing their workload and balancing their commitments. Students also get access to experienced career advisor Mike McCulloch, who can provide expert guidance.
A Supportive Environment
These services and resources create a supportive environment for OPIT students, which Karina says helped her throughout her course of study. Karina suggests organization and leaning into help from the community are the best ways to succeed when studying with OPIT.

Source:
- Agenda Digitale, published on May 16th, 2025
By Riccardo Ocleppo, Founder and Director of OPIT – Open Institute of Technology
AI ethics requires ongoing commitment. Organizations must integrate guidelines and a corporate culture geared towards responsibility and inclusiveness, preventing negative consequences for individuals and society.
In the world of artificial intelligence, concerns about algorithmic bias are coming to the forefront, calling for a collective effort to promote ethical practices in the development and use of AI.
This implies the need to understand the multiple causes and potential consequences of the biases themselves, identify concrete solutions and recognize the key role of academic institutions in this process.
Bias in AI is a form of injustice, often systemic, that can be embedded in algorithms. Its origins are many, but the main culprit is almost always the data set used to train the models. If this data reflects inequalities or prejudices present in society, the risk is that AI will absorb and reproduce them, consolidating these distortions.
But bias can also manifest itself in the opposite direction. This is what happened some time ago with Google Gemini. The generative AI system developed by Google, in an attempt to ensure greater inclusivity, ended up generating content and images completely disconnected from the reality it was supposed to represent.
Further complicating the picture is the very nature of AI models, which are often characterized by complex algorithms and opaque decision-making processes. This complexity makes it difficult to identify, and therefore correct, biases inherent in the systems.
Ethical Data Management to Reduce Bias in AI
Adopting good data management practices is essential to address these issues. The first step is to ensure that the datasets used for training are diverse and representative. This means actively seeking data that includes a wide variety of demographic, cultural, and social contexts, so as to avoid AI exclusively reproducing existing and potentially biased models.
Alongside data diversification, it is equally important to test models on different demographic groups. Only in this way can latent biases that would otherwise remain invisible be highlighted. Furthermore, promoting transparency in algorithms and decision-making processes is crucial. Transparency allows for critical control and makes all actors involved in the design and use of AI accountable.
Strategies for ethical and responsible artificial intelligence
Building ethical AI is not an isolated action, but an ongoing journey that requires constant attention and updating. This commitment is divided into several fundamental steps. First, ethical guidelines must be defined. Organizations must clearly establish the ethical standards to follow in the development and use of AI, inspired by fundamental values such as fairness, responsibility and transparency. These principles serve as a compass to guide all projects.
It is also essential to include a plurality of perspectives in the development of AI. Multidisciplinary teams, composed of technologists, ethicists, sociologists and representatives of the potentially involved communities, can help prevent and correct biases thanks to the variety of approaches. Last but not least, promote an ethical culture : in addition to establishing rules and composing diverse teams, it is essential to cultivate a corporate culture that places ethics at the center of every project. Only by integrating these values in the DNA of the organization can we ensure that ethics is a founding element of the development of AI.
The consequences of biased artificial intelligence
Ignoring the problem of bias can have serious and unpredictable consequences, with profound impacts on different areas of our lives. From the reinforcement of social inequalities to the loss of trust in AI-based systems, the risk is to fuel skepticism and resistance towards technological innovation. AI, if distorted, can negatively influence crucial decisions in sectors such as healthcare, employment and justice. Think, for example, of loan selection algorithms that unfairly penalize certain categories, or facial recognition software that incorrectly identifies people, with possible legal consequences. These are just some of the situations in which an unethical use of AI can worsen existing inequalities.
University training and research to counter bias in AI
Universities and higher education institutions have a crucial responsibility to address bias and promote ethical practices in AI development. Ethics must certainly be integrated into educational curricula. By including ethics modules in AI and computer science courses, universities can provide new generations of developers with the tools to recognize and address bias, contributing to more equitable and inclusive design. Universities can also be protagonists through research.
Academic institutions, with their autonomy and expertise, can explore the complexities of bias in depth, developing innovative solutions for detecting and mitigating bias. Since the topic of bias is multidimensional in nature, a collaborative approach is needed, thus fostering interdisciplinary collaboration. Universities can create spaces where computer scientists, ethicists, lawyers, and social scientists work together, offering more comprehensive and innovative solutions.
But that’s not all. As places of critical thinking and debate, universities can foster dialogue between developers, policy makers, and citizens through events, workshops, and conferences. This engagement is essential to raise awareness and promote responsible use of AI.
In this direction, several universities have already activated degree courses in artificial intelligence that combine advanced technical skills (in areas such as machine learning, computer vision and natural language processing) with training that is attentive to ethical and human implications.
Academic Opportunities for an Equitable AI Future
More and more universities around the world – including Yale and Oxford – are also creating research departments dedicated to AI and ethics.
The path to ethical AI is complex, but it also represents an opportunity to build a future where technology truly serves the common good.
By recognizing the root causes of bias , adopting responsible data practices, and engaging in ongoing and vigilant development, we can reduce the unintended effects of biased algorithms. In this process, academic institutions – thanks to their expertise and authority – are at the forefront, helping to shape a more equitable and inclusive digital age.
Read the full article below:
Have questions?
Visit our FAQ page or get in touch with us!
Write us at +39 335 576 0263
Get in touch at hello@opit.com
Talk to one of our Study Advisors
We are international
We can speak in: