View allAll Photos Tagged interface

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

For a while now an app called Ration has been floating around the MacTalk forums. It's a little app that tracks your download quota from Australian ISPs. Anyways, I whipped this up this arvo after looking at the current interface and imagining how much more streamlined it could be.

 

What do you think? :)

 

Ration: blargsoft.com/

Forwarded from Ross Koppel.

Waiting to exit the posession, 66107 Top & Tailed with 66160 stand on the Down Main with 6K10 Arbroath to Mossend "MOBC" while 66100 awaits with 6K11 Arbroath to Mossend "Engineers" on the Up Main.

Interface design for newsletter system

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

YouTube used to have an fun interface for exploring videos that could be launched in fullscreen directly from a YouTube video page. I don't know why, but it seems like this feature has been disabled. This is a screenshot from January 2008.

 

Videos are represented by circles, and if I remember correctly, new and related videos appeared when you clicked a specific video. In this screenshot a video is being played.

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The airodump-ng interface.

 

Airodump-ng can be used to scan for WiFi networks and capture the data packets.

San Rafeal, Mendoza, Argentina

Cañón del Atuel, Valle Grande

 

Visto desde el dique de Valle Grande, las cadenas montañosas semicubiertas por un velo de sombra que surge y muere en su misma naturaleza: la propia montaña.

El sol poniéndose, remolón, y una inmensa nube que se despierta, quizás llamada por la noche que ya empieza.

Se habló de lluvia, pero nada llovió... sólo algo de viento turbulento... sólo eso y el paisaje que todo lo vale.

In Interactive Design 3, students are asked to select from one of three client briefs to develop a social change project. Cherie, Michael and Jay chose to develop a mobile application to assist with earthquake preparedness. The application offers device features, such as RSS, GPS tracking and "bounce location" to sustain user correspondence during disaster relief efforts.

 

The team divided the roles and responsibilities to tackle strategic review, competitive analysis, moodboards, user experience, information architecture and interactive design.

 

Learn more about VFS's one-year Digital Design program at www.vfs.com/digitaldesign.

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

wireframe réaliser le 04 04 2011

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

Some people will know what this is. I'll post more pics of my "studio" later.

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

Interface observed on the street #2

a mock interface drawn up to help identify the various component that would need to be delivered to the end user for personalised delivery of resources.

 

Mock Outline for JISC Programme Strand on "Semantic Web Personalisation":

1.) Subject specific tool to pull contextual metadata from resources

1.1.) from human resources (subject curators)

1.1.1.) automated machine extraction of human profiles (contextual metadata)

1.1.2.) manual extraction human2human editorial process of discussion, question and answer

1.2.) from scholarly resources

2.) Cross Disciplinary Tools [didn't get the opportunity to discuss]

3.) Matching Algorythm

3.1.) human profile standards

3.1.1.) which standards should be used

3.1.1.1.) Standards from Web, e.g. FOAF, SIOC, hCard, hCalendar

3.1.1.2.) Subject specific standards

3.1.1.2.1.) which standards can be borrowed and from where?

3.1.1.2.2.) which ontologies will need to be created?

3.2.) knowledge artifact relations

3.2.1.) profile matching 1:N (heavy lifting)

3.2.1.1.) interface that is humanly understandable

3.2.2.) item profile match N:N (customized profiles)

3.2.2.1.) Investigatory

4.) Investigations on ontological drift (ontological crosswalks)

4.1.) Subject area specific change

4.1.1.) committee based human editorial board

4.1.1.1) publishing editorial process (dictionary)

4.1.2.) network dispersed teams

4.1.2.1.) developer plus subject expert mapping across

5.) Usability of human interfaces for interacting with tools

5.1) web based interfaces

5.2.) component interfaces for integration on 3rd party web sites

5.3.) API for utilization of tools

 

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

Button icons on the sprouts that aren't buttons and the order of top to bottom and left to right is the wrong way around.

The tissue processor is an automated machine that progressively dehydrates tissue and impregnates it with molten paraffin. This procedure usually runs overnight, although small biopsies can be processed in a few hours.

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

Employing a revolutionary technology developed at EPFL, Mental Work is the first showroom in the world to be operated exclusively by workers’ minds, juxtaposing the industrial revolution against the possibly imminent cognitive revolution — a symbiosis of human -and machine.

 

For the May 17, 2018 opening, swissnex San Francisco welcomed Mental Work co-founders, chief business officer Michael Mitchell, and chief scientific officer José Millán, (absent was chief visionary officer, Jonathon Keats) hosted a panel discussion of the scientific and conceptual elements of the showcase.

 

PHOTOS ASTRA BRINKMANN FOR SWISSNEX SF

Audiovisual Dance Performance. Presented in March 2015 at UT Austin – Oscar G. Brockett Theatre, Austin/TX.

 

video + info: visiophone-lab.com/wp/?portfolio=ad-mortuos

 

“Ad Mortuos” is a collaborative work inspired by poet Stephanie Pope, read by LaQuetta Carpenter.

 

Choreographer: Yacov Sharir;

Composer/Sound Design: Bruce Pennycook;

Visuals: João Data, Rodrigo Carvalho

Costume Design: Kelsey Vidic;

Vocalist: Yago de Quay;

Dancers: Emily Snouffer, Rebecca Bagley, Gianina Casale, Summer Fiaschetti, Katie McCarn and Ally Morales.

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

The media consumption experience is poised to transform, and fast. Technologies that have been tinkered with for years, ranging from virtual and augmented reality to sensors and robotics, are finally on the tipping point of mass commercialization. As the physical and digital worlds converge, how will these technologies shape how people interact with digital media?

 

On November 18, 2014, NYC Media Lab and Razorfish hosted the second occasion of Future Interfaces, an evening "science fair" on the future of human-computer interaction and digital media. More than 300 guests came to go hands-on with 30 demos from startups and universities to see what's on the verge of commercialization, what’s still in the lab, and what advances will change the nature of media and communications in the future.

 

To learn more about the event and to see a full list of participating demos, visit www.nycmedialab.org/events/future-interfaces/

YouTube used to have an fun interface for exploring videos that could be launched in fullscreen directly from a YouTube video page. I don't know why, but it seems like this feature has been disabled. This is a screenshot from January 2008.

 

Videos are represented by circles, and if I remember correctly, new and related videos appeared when you clicked a specific video.

Other than the doctored account number, this is a screen shot from my online water bill. Note that the payment history list is oldest first, while the billing history list is most-recent first.

 

I'm sure there's a reason for this. It's just not a good reason.

The GE-Zenon MBR process has a stand-alone operator interface.

view at: bit.ly/xutxBc

Hi there! This is my new project: an Hi-tech Touch interface project! Use it to show your portfolio in a hi-tech way!§

This project has a modular structure : there are two modules of 6 placeholders that you can duplicate to add all photos/videos needed. The native project template comes with 12 placeholders for your photos/video. But There is also a version with 24 palceholders (bonus version!). Just import your photos, type your texts and you’ll be ready to render! In the help file I explain how to add more photos.

Main features:

 

HD native resolution (1280×720). 25 fps.

CS4 project. CS5 compatible.

1 min 30 sec length of native version (with 12 placeholders).

2 min 40 sec length of bonus version (with 24 placeholders).

No plugins required.

Easy to customize.

Step-by-step instructions included (pdf file) to change texts, video/photo, colors, add photo.

Free font used (txt with link).

Music and effects used in the preview are not included.

 

Enjoy!

Music track used in the preview is our bit.ly/wDI5mS song.

Sound Effects are made by bit.ly/AtZtKF and they’re available here : bit.ly/zB9i5X .

Please note that inside his “Electronic sound Pack” you’ll find also a track with all sound effects syncronized for this template, so you won’t have to arrange single effects inside the template (version with 12 placeholders).

Now these fit on a USB plug.

1 2 ••• 5 6 8 10 11 ••• 79 80