ICC TOKYO 2016

ICC (Industry Co-Creation) Tokyo 2016にて、暦本がパネリストとして登壇しました。

2016年3月24日開催 ICCカンファレンス TOKYO 2016
最先端研究の動向(人工知能 コグニティブ IoA)
武田 浩一   日本アイ・ビー・エム株式会社 技術理事
松尾 豊    東京大学大学院 特任准教授
暦本 純一   東京大学大学院情報学環 教授/ソニーコンピュータサイエンス研究所副所長
(モデレーター)
田川 欣哉     takram design engineering 代表

Part 1: 「IoTからIoAへ」最先端の研究者が考える人工知能と人間の未来」

Part 2:「人間の仕事はAI(人工知能)に代替されてしまうのか?」最先端の研究者が描く知能と生命の境界

Part 3: 「人工知能時代に磨くべき力とは?」最先端の研究者が語る今後求められるスキル」

情報学環/総合分析情報学コース入試説明会

2016年6月5日(日)に、暦本研究室が所属しています東京大学情報学環および総合分析情報学コースの入試説明会があります。受験を検討されている方はふるってご参加ください:

東京大学 学際情報学府 学際情報学専攻
総合分析情報学コース 第二回 コース説明会2016
2016年6月5日(日)10:00〜11:30
場所:本郷キャンパス(〒113-0033 東京都文京区本郷7-3-1)
ダイワユビキタス学術研究館 3階 ダイワハウス石橋信夫記念ホール

平成29年度東京大学大学院学際情報学府入試説明会
Orientation for the entrance examination of the graduate school of interdisciplinary information studies, the University of Tokyo
2016年6月5日(日)13:00 – 17:00
Sunday, June 5, 2016, 13:00-17:00
場所:東京大学本郷キャンパス・情報学環福武ホール地下2階ラーニングシアター
http://fukutake.iii.u-tokyo.ac.jp/access/
B2 floor, Fukutake Hall, Hongo Campus

どちらも申込不要、入場無料です。

 

ACE 2015

The following papers will be presented at ACE 2015 (Advances in Computer Entertainment Technology 2015), Iskandar, Malaysia (Nov.16-19, 2015).

  • “MoveMe: 3D Haptic Support for a Musical Instrument”, Katsuya Fujii, Sophia Russo, Pattie Maes and Jun Rekimoto
  • “Shepherd Pass: Ability Tuning for Augmented Sports using Ball-Shaped Quadcopter”, Kei Nitta, Keita Higuchi, Yuichi Tadokoro and Jun Rekimoto

HCI talks by SIGCHI Researchers (ユビキタス情報社会基盤研究センター共催)

After SIGCHI conference in Seoul, two HCI researchers will visit Tokyo University and give talks, open for everybody.

Date: April 27 (mon) 13:30-15:00 (door open at 13:15)
Venue: Ishibashi Hall, 3F, Daiwa Ubiquitous Computing Research Building
http://www.u-tokyo.ac.jp/content/400020145.pdf (building #125)

TALK 1

Alex Olwal
Ph.D, Google, USA
http://www.olwal.com

TITLE: Augmented Realism through Relevant Minimalism

ABSTRACT
Augmented Reality (AR) merges virtual information with a real environment for intuitive and direct user interfaces. I will give an overview of our research projects that emphasize a seamless blend of the digital and physical. We leverage exotic displays, sensing and context, to register digital content with the environment and to support rich interactions. Our goal is to render less, with every pixel being relevant, dynamic and grounded in the space. This is one of our key philosophies for enhancing the human senses and capabilities through minimal, yet highly relevant augmentations, that embrace the realism of our physical world. Our projects include novel interaction techniques, see-through displays, sensing technologies, immaterial user interfaces, and dynamic shape displays.

BIO
Alex Olwal (Ph.D., M.Sc.) is an Interaction Researcher at Google, Affiliate Faculty at KTH, and Research Affiliate at the MIT Media Lab.

Alex designs and develops interactions and technologies that embrace digital and physical experiences. He is interested in tools, techniques and devices that enable new interaction concepts for the augmentation and empowerment of the human senses.

Alex’s research (olwal.com) includes augmented reality, spatially aware mobile devices, medical user interfaces, ubiquitous computing, touch-screens, as well as novel interaction devices and displays.

He has previously worked with the development of new technologies for Human-Computer Interaction at MIT – Massachusetts Institute of Technology (Cambridge, MA), KTH – Royal Institute of Technology (Stockholm), Columbia University (NY), University of California (Santa Barbara, CA) and Microsoft Research (Redmond, WA).

At Google, Alex conducts applied research and development of novel and exotic input and output mechanisms for wearables, and explore associated
interaction techniques. Alex’s goal is to expand expressiveness while avoiding interference with user’s experience of the physical reality.

TALK 2:

Pedro Lopes
Ph.D Candidate, Computer Interaction Lab, Hasso Platner Institute, Germany
http://plopes.org/

TITLE: Affordance++ and Proprioceptive Interaction

ABSTRACT
We propose extending the affordance of objects by allowing them to communicate dynamic use, such as (1) motion (e.g., spray can shakes when touched), (2) multi-step processes (e.g., spray can sprays only after shaking), and (3) behaviors that change over time (e.g., empty spray can does not allow spraying anymore). Rather than enhancing objects directly, however, we implement this concept by enhancing the user. We call this affordance++. By stimulating the user’s arms using electrical muscle stimulation, our prototype allows objects not only to make the user actuate them, but also perform required movements while merely approaching the object, such as not to touch objects that do not “want” to be touched. In our user study, affordance++ helped participants to successfully operate devices of poor natural affordance, such as a multi-functional slicer tool or a magnetic nail sweeper, and to stay away from cups filled with hot liquids. We call this concept of creating object behavior by controlling user behavior affordance++. Conceptually there are many ways of implementing affordance++, generally by applying sensors and actuators to the user’s body, such as the arm. We actuate users by controlling their arm poses using electrical muscle stimulation, i.e., users wear a device on their arm that talks to the user’s muscles by means of electrodes attached to the user’s arm. This allows for a particularly compact form factor and is arguably even more “direct” than the indirection through a mechanical system. However, the concept of affordance++ needs not to be tied to a particular means of actuating the user, but to the concept of doing so instead of actuating the objects that the user interacts with.

BIO
http://plopes.org/bio/

CHI 2015 Presentations & Activities

At CHI2015, The following papers will be presented from our group.

  • ChameleonMask: Embodied Physical and Social Telepresence using human surrogates
    Authors
    , Kana Misawa and Jun Rekimoto (Monday, April 20, Time: 11:30 – 12:50, room 308, alt.chi “Augmentation”)

  • ImmerseBoard: Immersive Telepresence Experience using a Digital Whiteboard
    Keita Higuchi, Yinpeng Chen, Philip A Chou, Zhengyou Zhang, Zicheng Liu (Wednesday, April 22, 2015
    Time: 9:30 – 10:50, papers “Telepresence Video, Robots, and Walls”, room E6)