Showing posts with label scholars. Show all posts
Showing posts with label scholars. Show all posts

Tuesday, June 27, 2023

A Dishonesty Expert Stands Accused of Fraud. Scholars Who Worked With Her Are Scrambling.; The Chronicle of Higher Education, June 22, 2023

Nell Gluckman
, The Chronicle of Higher Education; A Dishonesty Expert Stands Accused of Fraud. Scholars Who Worked With Her Are Scrambling.

"To Maurice Schweitzer, a University of Pennsylvania professor, it seemed logical to team up with Francesca Gino, a rising star at Harvard Business School. They were both fascinated by the unseemly side of human behavior — misleading, cheating, lying in order to profit — and together, they published eight studies over nearly a decade.

Now, Schweitzer wonders if he was the one being deceived."

Sunday, April 10, 2022

AI.Humanity Ethics Lecture Series will explore the ethics of artificial intelligence; Emory University, Emory News Center, April 5, 2022

Emory University, Emory News Center; AI.Humanity Ethics Lecture Series will explore the ethics of artificial intelligence

"As society increasingly relies on artificial intelligence (AI) technologies, how can ethically committed individuals and institutions articulate values to guide their development and respond to emerging problems? Join the Office of the Provost to explore the ethical implications of AI in a new AI.Humanity Ethics Lecture Series.

Over four weeks in April and May, world-renowned AI scholars will visit Emory to discuss the moral and social complexities of AI and how it may be shaped for the benefit of humanity. A reception will follow each lecture.

Matthias Scheutz: “Moral Robots? How to Make AI Agents Fit for Human Societies” 

Monday, April 11

Lecture at 4 p.m., reception at 5:30 p.m.

Convocation Hall — Community Room (210)

Register here.

AI is different from other technologies in that it enables and creates machines that can perceive the world and act on it autonomously. We are on the verge of creating sentient machines that could significantly improve our lives and better human societies. Yet AI also poses dangers that are ours to mitigate. In this presentation, Scheutz will argue that AI-enabled systems — in particular, autonomous robots — must have moral competence: they need to be aware of human social and moral norms, be able to follow these norms and justify their decisions in ways that humans understand. Throughout the presentation, Scheutz will give examples from his work on AI robots and human-robot interaction to demonstrate a vision for ethical autonomous robots...

Seth Lazar: “The Nature and Justification of Algorithmic Power” 

Monday, April 18

Lecture at 4 p.m., reception at 5:30 p.m.

Convocation Hall — Community Room (210)

Register here.

Algorithms increasingly mediate and govern our social relations. In doing so, they exercise a distinct kind of intermediary power: they exercise power over us; they shape power relations between us; and they shape our overarching social structures. Sometimes, when new forms of power emerge, our task is simply to eliminate them. However, algorithmic intermediaries can enable new kinds of human flourishing and could advance social structures that are otherwise resistant to progress. Our task, then, is to understand and diagnose algorithmic power and determine whether and how it can be justified. In this lecture, Lazar will propose a framework to guide our efforts, with particular attention to the conditions under which private algorithmic power either can, or must not, be tolerated.

Ifeoma Ajunwa: “The Unrealized Promise of Artificial Intelligence” 

Thursday, April 28

Lecture at 4 p.m., reception at 5:30 p.m.

Oxford Road Building — Presentation Room and Living Room/Patio

Register here.

AI was forecast to revolutionize the world for the better. Yet this promise is still unrealized. Instead, there is a growing mountain of evidence that automated decision making is not revolutionary; rather, it has tended to replicate the status quo, including the biases embedded in our societal systems. The question, then, is what can be done? The answer is twofold: One part looks to what can be done to prevent the reality of automated decision making both enabling and obscuring human bias. The second looks toward proactive measures that could allow AI to work for the greater good...

Carissa Véliz: “On Privacy and Self-Presentation Online” 

Thursday, May 5

Lecture at 4 p.m. 

Online via Zoom 

A long tradition in philosophy and sociology considers self-presentation as the main reason why privacy is valuable, often equating control over self-presentation and privacy. Véliz argues that, even though control over self-presentation and privacy are tightly connected, they are not the same — and overvaluing self-presentation leads us to misunderstand the threat to privacy online. Véliz argues that to combat some of the negative trends we witness online, we need, on the one hand, to cultivate a culture of privacy, in contrast to a culture of exposure (for example, the pressure on social media to be on display at all times). On the other hand, we need to readjust how we understand self-presentation  online."

Tuesday, August 30, 2016

Feds Target 'Predatory' Publishers; Inside Higher Ed, 8/29/16

Carl Straumsheim, Inside Higher Ed; Feds Target 'Predatory' Publishers:
"The Federal Trade Commission on Friday filed a complaint against the academic journal publisher OMICS Group and two of its subsidiaries, saying the publisher deceives scholars and misrepresents the editorial rigor of its journals.
The complaint, filed in the U.S. District Court for the District of Nevada, marks the first time the FTC has gone after what are often known as “predatory” publishers. Such publishers exploit open-access publishing as a way to charge steep fees to researchers who believe their work will be printed in legitimate journals, when in fact the journals may publish anyone who pays and lack even a basic peer-review process."

Friday, August 12, 2016

Think Tanks and the Influence of Corporate Dollars; New York Times, 8/10/16

Room for Debate, New York Times; Think Tanks and the Influence of Corporate Dollars:
"Think tanks inform both government policy and media analysis with their research, because as nonprofit institutions, they are seen as independent. But some institutions vigorously push their donors’ agendas, acting like lobbyists. Some scholars even use their positions at think tanks to promote work they are separately paid to do for corporations.
What can be done to protect against corporate influence over research institutions?"

Wednesday, December 2, 2015

Scholars Unveil New Edition of Hitler’s ‘Mein Kampf’; New York Times, 12/1/15

Alison Smale, New York Times; Scholars Unveil New Edition of Hitler’s ‘Mein Kampf’ :
"Not since 1945, when the Allies banned the dubious work and awarded the rights to the state of Bavaria, has Hitler’s manifesto, “Mein Kampf,” been officially published in German. Bavaria had refused to release it. But under German law, its copyright expires Dec. 31, the 70th year after the author’s death.
That allows a team of historians from a noted center for the study of Nazism, the Institute for Contemporary History in Munich, to publish its two-volume, 2,000-page edition, a three-year labor complete with about 3,500 academic annotations.
The intention is to set the work in historical context, to show how Hitler wove truth with half-truth and outright lie, and thus to defang any propagandistic effect while revealing Nazism."