This week at school we’ve been brainstorming and discussing qualities we want to see in our graduates as they leave our school and enter the wider world of college and work. Today among many other questions, I was thinking about automation and the ways we need students to develop computational thinking skills so they can effectively utilize mind-blowingly powerful AI’s (artificial intelligence bots) to collaborate with others and solve problems. As I gaze into the palantir afforded me as a co-host of “The EdTech Situation Room” each week with Jason Neiffer (@techsavvyteach), this is part of the future I see for our students, our society, and ourselves in the coming decades.

This 98 second video, which I titled “EdTech Situation Room Promo Trailer,” is the result of my thinking about this question tonight. Spoiler: The video offers more questions than answers, and as an answer suggests you need to follow @edtechSR on Twitter and subscribe to our weekly webshow and podcast on YouTube and on edtechSR.com!

 

This question of what an emerging “artificial intelligence first” rather than “mobile first” worldview (which Google announced at Google IO 2017) should mean for schools is something I discussed on The EdTech Situation Room back on May 17, 2017, with Jason Neiffer (@techsavvyteach) and Ben Wilkoff (@bhwilkoff). It’s not something I will claim to have definitively figured out, but it’s definitely a question with which I’m likely continue to wrestling in the months ahead.

I created this short video by:

  1. Writing a script in Google Docs.
  2. Finding and saving related images for each script sentence on PhotosForClass.com, thenounproject.com, and WikiPedia.
  3. Saving SVG files from The Noun Project into Documents 6 for iPad since the PNGs wouldn’t directly download or save to my iPad. I then uploaded those to Google Drive from the app.
  4. Converting Noun Project SVG image files to medium resolution JPG image files using cloudconvert.com (free) and Google Drive with my laptop.
  5. AirDropping converted images from my laptop to my iPad.
  6. Importing images into Adobe Spark Video for iPad.
  7. Expressively recording my script audio on each image in Adobe Spark Video.
  8. Selecting an appropriate musical track for the background within the Adobe Spark Video app, and slightly reducing the volume.
  9. Exporting the finished video to my iPad camera roll.
  10. Uploading the final video to my YouTube channel with the YouTube iPad app.

Can you tell I love digital storytelling and acting a bit “wack-a-doo” (to borrow a favorite word from my wife) when it comes to audio recording?! If you haven’t given Adobe Spark Video a try yourself, you definitely should. It supports a streamlined workflow for creating narrated slideshows and digital stories that’s worth using personally as well as in the classroom!

Check out the “Narrated Slideshow – Screencast” and “Digital Storytelling” pages of ShowWithMedia.com for additional resources and examples related to these media project types.


Did you know Wes has published several eBooks and "eBook singles?" 1 of them is available free! Check them out!

Do you use a smartphone or tablet? Subscribe to Wes' free magazine "iReading" on Flipboard!


If you're trying to listen to a podcast episode and it's not working, check this status page. (Wes is migrating his podcasts to Amazon S3 for hosting.) Remember to follow Wesley Fryer on Twitter (@wfryer), Facebook and Google+. Also "like" Wesley's Facebook pages for "Speed of Creativity Learning" and his eBook, "Playing with Media." Don't miss Wesley's latest technology integration project, "Mapping Media to the Curriculum."

On this day..

Tagged with →  
Share →
Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 Unported License.

Made with Love in Oklahoma City