Do I need this tool? Why? How does it really support learning? What are the costs, both monetary and otherwise, of using this service? Do the rewards of use outweigh the risks? Is there a paid service I could explore that will meet my needs and better protect the privacy of my information and my students’ information? How can I inform parents/community members about our use of this tool and what mechanisms are in place for parents to opt their children out of using it? When this tool and/or its plan changes, how will we adjust? What will our plans be to make seamless transitions to other tools or strategies when the inevitable happens?
There has been a quiet push lately by tech industry giants to get ethical about future technologies. But is anything more than PR? And how do we teach technology students to preempt a possible ethical disaster? Jordan Erica Webber explores the issues
Watch Dr Simon Longstaff’s presentation for more on ethics.
23:21 Larry Prusak: Knowledge and it’s Practices in the 21st Century
Prusak discusses the changes in knowledge over time and the impact that this has. This reminds me of Weinberger’s book Too Big To Know. Some quotes that stood out were:
Knowledge won’t flow without trust
Schools measure things they can measure even if it is not valuable
Again and again Prusak talks about going wide, getting out and meeting new people.
1:21:59 Professor Genevieve Bell: Being Human in a Digital Age
Bell points out that computing has become about the creation, circulation, curation and resistence of data. All companies are data companies now. For example, Westfield used to be a real estate company, but they are now a data company.
The problem with algorithms is that they are based on the familiar and retrospective, they do not account for wonder and serendipity.
As we design and develop standards for tomorrow, we need to think about the diversity associated with those boards and committees. If there are only white males at the table, how does this account for other perspectives.
We do want to be disconnected, even if Silicon Valley is built around being permanently connected. One of the things that we need to consider is what is means to have an analogue footprint.
Building on the discussion of data and trust, Bell makes the point:
The thing about trust is that you only get it once.
The question remains, who do we trust when our smart devices start selling our data.
In regards to the rise of the robots, our concern should be the artificial intelligence within them. One of the big problems is that robots follow rules and we don’t.
The future of technology that we need to be aspiring to develop a future where technology can support us with our art, wonder and curiosity.
A comment made during the presentation and shared after Bell had finished:
Is your current job the best place for you to make the world a better place?
2:49:51 Phillip Bernstein: The Future of Making Things: Design Practice in the Era of Connected Technology
Berstein unpacks six technical disruptions – data, computational design, simulation analysis, the internet of things, industrial construction and machine learning – and looks at the implications for architecture.
3:51:44 Dr Simon Longstaff: Ethics in the Next Machine Age
Dr Longstaff explores the ethics associated with technology. This includes the consideration of ethical design, a future vision – Athens or Eden – and the purpose to making. Discussing the technology of WWII, Longstaff states:
Technical mastery devoid of ethics is the root of all evil
He notes that just because we can, it does not mean we ought.
He also used two ads from AOL to contrast the choices for tomorrow:
H/T Tom Barrett