Whitman website designers
Hello and welcome to this website designers Web Designer Whitman video tutorial.
I’m Owen Corso from Google.
And today, we’re going to build a rich media expandable creative with video.
Let’s start by selecting file, New File.
This opens a dialog box where we will set up our ad.
First, let’s make out high of project.
We have four options– The default is Display & Video 360so we will leave that as is.
How to Be Comfortable in the Dentist's Chair
KENT: My name is Kent.
I'm the creativespecialist on the Google Web Designer team and I'd like to review dynamic exits in Studio A dynamic exit refers to the URL loaded when a dynamic ad is clicked on It's usually related to the product being shown and that URL comes from yourdata feed This video describes the workflow for setting up dynamic exits in Studio Now the simplest case is an ad which doesn't use a gallery or carouselcomponent Here's an example from another one of my videos titled Dynamic inStudio - Google Web Designer You'll see my creative contains only placeholder content But when I preview it in Studio you'll see it loading data from my feed Refresh it Now we're seeing product number 0 Look at my feed for product 0 we're seeing this name, this image and here is the URLfor my exit for this product Now when I set up this ad, I used the simple static exit So when I click anywhere on the ad it'll load a static URL like a landingpage What we want to have happen is this to exit to the URL associated for that product Let me show you how to do that in Google Web Designer First thing we'll do is select the current event and delete it Then we're going to look over inour components panel and open up Interaction and grab a taparea Then with the Transform control selected I'm going to stretch this out to cover the ad And in my Dynamic panel I'm going to click the plus icon to make a new binding We'll select that tap area We're going to look for Exit override URL And we'regoing to drill into our feed the first item, exit URL, and get the URL Click OK and we'llpublish this Now when that's done we'll switch over toStudio We'll reload our creative I'm going to click on the ad Now I'll see we'regetting product one Just to make sure this is really working let's reload itand get another image Here's number five Let's click on that Here's product fiveso that's good Now next thing you might want to do is combine static and dynamicexits in a single ad so you might want to have a logo up here which goes to a home page and maybe some disclaimer text down here which goes to a legal page So let's see how we can do that Back in Google Web Designer I'm going to double click this newtaparea and give it a new name exit-default then I'll copy and paste it I'm going to call that one exit-product and this layer is on top So I'm going tomake this just cover the product area and I'm going to select exit-default andchange the exit on that Now one trick over here in the Dynamic panel I can choose Selection and now I would just see the binding on that selection I'm going to select it and delete it With it still selected I'm going to move to the Events panel click the plus icon and add a new event Google ad, exit ad, gwd-ad I'm going to put an arbitrary ID in here and I'm going to put a full URL to mylanding page OK and we're going to publish again Now when that's done, we'll switch back to Studio and reload the page Now when I click in the corners I'm getting my static URL And when I click on the product I'm getting the correct product That concludes this demo, dynamic exits in Studio Thanks for watching.
Whitman website designers
Next, we can select the type of ad.
We want to make an expandable, so we select Expandable on the left.
Next, we can set again ad’s dimensions.
We are building a 320 by 50that expands to 480 by 250.
So I will make those changes.
We then assign the Whitman creative a name.
I will leave my Save ToLocation as the default, and leave the talk about set to Quick.
Once I’m happy with my settings, I click OK.
Google Web Designer creates the initial pages of the ad for me with the dimensions I defined.
The collapsed page already contains a Tap Area event to expand the ad and an expanded pageWhitman with a close tap area to collapse back down.
From product design to virtual reality
I am a product designer at Google, and I joined the company through Sparrow, a French startup that got acquired on July 20, 2012. Since then, I worked with the Gmail team to build from scratch a flagship product that became Inbox by Gmail. It shipped on October 22, 2014.
I designed productive applications for a few years, and I felt like I reached a tipping point. I wanted to expand my skill set, learn new things every day and get better at something I’ve never touched. I needed new challenges to reboot myself by leaving my comfort zone.
I got interested in virtual reality around the Oculus Kickstarter period because of the immersiveness and endless possibilities that came with it. There is nothing more exciting than building for a new medium and exploring an uncharted territory.
I joined the Google Cardboard and Virtual Reality team on April 17, 2015. Thanks to Clay Bavor and Jon Wiley for this great opportunity.
My first weeks in the team were as scary as it can get. People used words I had no idea of and asked me questions I didn’t know how to answer.
I am not going to lie, ramping up on the jargon was not easy but I was expecting that. Virtual reality is a deep field (pun intended) grouping together a variety of job titles each requiring a very specialized skill set. The first weeks were intense and day after day I had a better vision of the big picture. Slowly, the pieces came together. I found out which roles would be the best fit, what I wanted to do and what was required to get there. Regardless of the mission, I knew I would have to learn a lot, but I was prepared for this challenge. My feelings varied from one day to another. From super excited to create and learn something new, to super scared because of the colossal knowledge I still had to learn. Working with smart and knowledgeable people around me reinforced these mixed feelings.
Everything is going to be alright
I told myself and firmly believed that the dots would connect eventually. I am a passionate person, and I knew that I didn’t mind spending hours learning and experimenting.
During my product designer career, I got better at understanding, identifying and resolving user problems. Making things easy to use and delight users is not that different, no matter the medium.
The core of the mission is the same, but to get you from point A to B there are some interesting things to know.
- Sketching, is, again, at the core of everything. During any brain dump or design phase, sketching is as fast as it can get. I’ve sketched more in the time I’ve joined this team than I have in my entire career.
- Any design skills as diverse as they are will be a huge benefit.
- Photography knowledge will help you because you will interact with concepts such as field of view, depth of field, caustics, exposure and so on. Being able to use light to your advantage has been much valuable to me already.
- The more you know 3D and tools, the less you will have to learn. It’s pretty obvious but be aware that at some point, you might do architecture, character, props modeling, rigging, UV mapping, texturing, dynamics, particles and so on.
- Motion design is important. As designers, we know how to work with devices with physical boundaries. VR has none, so it’s a different way of thinking. “How does this element appear and disappear?” will be a redundant question.
- Python, C#, C++ or any previous coding skills will help you ramp up faster. Prototyping has a big place because of the fundamental need of iterating. This area is so new that you might be one of the first to design a unique kind of interaction. Any recent game engine such as Unity or Unreal engine largely integrates code. There is a large active community in game and VR development with a huge amount of training and resources already.
- Be prepared to be scared and get ready to embrace the unknown. It’s a new world that evolves every day. Even the biggest industry-leading companies are still trying to figure things out. That’s how it is.
Design teams will evolve because this new medium opens a lot of possibilities for creation. Think about the video game or the film industry for instance.
I think there will be two big design buckets.
The first one will be about the core user experience, interface, and interaction design. This is very close to how product design team are structured today (Visual, UI, UX, motion designers, researchers, and prototypers).
Each role will have to adapt to the rules of this new medium and keep a tight relationship with engineers. The goal will always remain the same; create a fast iteration cycle to explore a wide range of interactive designs.
On the other hand, content teams will replicate indie and game design studio structure to create everything from unique experiences to AAA games. The entertainment industry as we know it in other mediums will likely be very similar in VR.
Ultimately, both will have a close relationship to create a premium end to end experience. Both industries have a great opportunity to learn from each other.
To wrap up on my personal experience, I think being a product designer in VR is not that different but requires a lot of dedication to understand and learn a vast field of knowledge.
First step and fundamentals of VR design
In this second part of the article, I will try to cover the basics you need to know regarding this medium. It’s meant to be designer oriented and simplified as much as possible.
Let’s get (a little bit) technical
The new dimension and immersiveness is a game changer. There is a set of intrinsic rules you need to know to be able to respect physiologically and treat your users carefully. We regrouped some of these principles in an app so you can learn through this great immersive experience.
Download Cardboard Design Lab
You can watch Alex’s presentation at I/O this year which goes more in-depth. The following is a small summary.
If you have to remember just two rules:
- Do not drop frames.
- Maintain head tracking.
People instinctively react to external events you might not be aware of, and you should be designing accordingly.
Physiological comfort. It regroups notions like motion sickness. Be careful when using acceleration and deceleration. Maintain a stable horizon line to avoid the “sea sickness” effect.
Environment comfort. People can experience various discomforts in certain situation like heights, small spaces (claustrophobia), big spaces (agoraphobia) and so on. Be careful with the scale and colliding objects. For example, if someone throws an object at you, you will instinctively try to grab it, dodge or protect yourself. Use it to your advantage and not to user’s disadvantage.
You can also use user senses to help you create more immersive products and cues. You can find inspiration in the game industry. They use all sorts of tricks to guide users during their journey. Here’s a couple:
- Audio for spatial positioning.
- Light to show a path and help the player.
Do not hurt or over-fatigue your users. It’s a classic mistake when you start to design for this medium. As cool as it looks, Hollywood sci-fi movies fed us with interactions that goes against simple ergonomic rules and can create major discomfort over time. Minority Report gestures are not suitable for a long period of activity.
I made a very simplified illustration of XY head movement safe zones. Green is good, yellow is ok and avoid red. There are a some user studies made public (links at the bottom of the page) that will give you more in-depth information about that topic.A simplified illustration of XY head movement safe zones.
Bad design can lead to more serious conditions.
As an example, have you heard about Text Neck? A study, published in Neuro and Spine Surgery measured varying pressures in our neck as our head moves to different positions. Moving from a neutral head position looking straight ahead to looking down increase the pressure by 440%. The muscles and ligaments get tired and sore; the nerves are stretched, and discs get compressed. All of this misbehavior can lead to serious long-term issues such as permanent nerves damages.
TLDR; Avoid extended look down interactions.
Degrees of freedom
The body has six different ways of moving in space. It can rotate and translate in XYZ.
3 Degrees of freedom (Orientation tracking)
Phone based head mounted device such as Cardboard, Gear VR are tracking the orientation via an embedded gyroscope (3DOF). Rotations on all three axes are tracked.
6 Degrees of freedom (Orientation + Position tracking)
To achieve six degrees of freedom, the sensor(s) will track positions in space (+X, -X, +Y, -Y, +Z, -Z). High-end devices like HTC Vive or Oculus Rift are 6DOF.
Making 6DOF possible frequently involves optical tracking of infrared emitters by one or more sensors. In Oculus’s case, the tracking sensor is on a stationary camera, while in Vive’s case the tracking sensors are on the actual HMD.
Depending on the system you are designing for, the input method will vary and affect your decisions. For example Google Cardboard has a single button, that’s why the interaction model is a simple gaze and tap. HTC Vive uses two, six degrees of freedom controllers and Oculus will ship with an Xbox One controller but will eventually use a 6DOF dual controller “Oculus Touch”. All of them allow you to use more advanced and immersive interaction patterns.The good old Xbox OneOculus Touch
There are also other kinds of inputs such as hand tracking. The most famous being Leap Motion. You can mount it to your Head Mounted Display (HMD).Leap Motion on top of a DK2
This area constantly evolves as technology catches up but as of today, hand tracking is not reliable enough to be used as the main input. The principal issues are related to hands and fingers, collisions, and subtle movements tracking.
Even though it’s very familiar, using a game controller is a disappointing experience. It physically removes some of the freedom VR is creating. In FPS, strafing and moving might usually cause some discomfort because of the accelerations.
On the other hand, the HTC Vive controllers reinforce the VR experience thanks to the 6 degrees of freedom, and Tilt Brush is a really good example. As I am writing theses lines, I haven't tried the Oculus touch but every demo I have seen looks very promising. Check out Oculus Toybox demo.
While designing user interfaces and interactions, inputs are the keystone that will drive some decisions differently depending on which method you are using. You should be familiar with all of them and aware of their limitations.
This is a big piece and might require a more in-depth article. I will focus on the most popular tools used in this industry.
Pen and paper
You just can’t beat them. It’s the first tool we use because it’s always around and does not require too many skills. It’s a proven way to express your ideas and iterate at a fast and cheap pace. Theses factors are important because, in VR, the cost of moving from wireframes to hi-fi is higher than 2D.
I still use it every day. Because of its ease of use, it’s the perfect tool that allows me to create a lot of explorations before moving to a VR prototype. It’s also handy for its export tools and plugins that are a huge time saver. If you are not familiar with that program, I wrote articles here and there.
I don’t see C4D as a competitor of Maya. Both are great tools, and each excels in its own way. When you don’t have a 3D background, the learning curve can be very steep. I like C4D because the interface, the parametric and non-destructive approach make sense for me. It helps me create more iterations quickly. I love the MoGraph modules, and a lot of great plugins are available. The community is very active, and you can find a lot of high-quality learning materials.Cinema 4D motion explorations
Maya is colossal in a good and a bad way. It does anything and everything a 3D artist needs. Most of the games and movies are designed with it. It’s a robust piece of software which can handle massive simulations and very heavy scenes with ease. From rendering, modeling, animation, rigging, it’s simply the best tool out there. Maya is highly customizable, and that is one reason why it’s the industry standard. Studios need to create their own set of tools, and Maya is the perfect candidate to integrate any pipeline.
On the other hand, learning all the tools will require your full and unconditional dedication for quite some time. I mean weeks of explorations, months of learning and years of practice on a daily basis.
It’s most certainly THE prototyping tool where everything will happen. You can easily create and move things around with a direct VR preview of your project. It’s a powerful game engine with a great community and a ton of resources available in their store (the asset author determines the pricing). In the assets library, you can find simple 3D models, complete projects, audio, analytics tools, shaders, scripts, materials, textures and so on.
Their documentation and learning platform are stellar. They have a wide range of high-quality tutorials.
It support all major HMD and is the best to build for cross-platform: Windows PC, Mac OS X, Linux, Web Player, WebGL, VR(including Hololens), SteamOS, iOS, Android, Windows Phone 8, Tizen, Android TV and Samsung SMART TV, as well as Xbox One & 360, PS4, Playstation Vita, and Wii U
It supports all major 3D formats and has the best in 2D game creation. The in-app 3D editor is weak, but people have built great plugins to correct that. The software is licensed based, but you can also use the free version to a certain extent. You can check the details on their pricing page. It’s the most popular game engine out there with ~47% of market share.
The direct competitor of Unity3D. Unreal also has great documentation and videos tutorials. Their store is smaller because it’s much newer.
One of the big advantages over the competition are the graphic capabilities; Unreal is one step ahead in nearly every area: terrain, particles, post processing effects, shadows & lighting, and shaders. Everything looks amazing.
Unreal Engine 4 uses C++ and comes with Blueprint, a visual script editor.
I haven’t worked with it too much yet, so I can’t elaborate more.
Less cross-platform compatibility: Windows PC, Mac OS X, iOS, Android, VR, Linux, SteamOS, HTML5, Xbox One, and PS4.
Virtual reality is a very young medium. As pioneers, we still have a lot to learn and discover. That’s why I am very excited about it and why I joined this team. We have the opportunity to explore and we should, as much as we can. Understand, identify, build and iterate. Over and over.
And over again…
- Immersive design Facebook group
- Google I/O 2015 — Designing for Virtual Reality
- Oculus Connect keynotes
- VR Design: Transitioning from a 2D to 3D Design Paradigm
- VR Interface Design Pre-Visualisation Methods
- 2014 Oculus Connect — Introduction to Audio in VR
- Cinema 4D tutorials
- Unity 3D tutorials
- Maya and 3D tools tutorials
- LeapMotion — VR Best Practices Guidelines
- The fundamentals of user experience in virtual reality
- Ready for UX in 3D?
Thanks everyone who helped me with the rereading and improvements 💖
Material Design Components for web - Designer vs. Developer #22
Ant (GitHub) is much more than a React UI kit with a minimalist design aesthetic and every component under the sun. It is a rabbit’s hole that leads to a giant maze of interconnected libraries, with a serious ecosystem surrounding it. There’s a custom build tool based on Webpack called ant-tool, several CLI apps, community scaffolds, and a complete framework (dva, which has its own CLI as well). And the UI components are mini-projects in and of themselves — see this repo for information on each component.
Many of these libraries appear to be very polished, including an entire React animation library. And I’d love to learn more about them, but Ant comes with a challenge — the majority of the documentation is in Chinese.
How’s Your Chinese?
Let me preface this by pointing out that the components library and its terrific style guide have been translated into English by generous volunteers, so the UI kit is completely usable. And the translation effort demonstrates the project’s intentions to open up Ant to a wider audience, boding well for companies considering adopting it.
However, there are some language issues that remain. The English is sometimes confusing or obscure. The maintainer of the library has commented here that they welcome PRs for improving the documentation, so that could be a great way to get involved in this amazing project.Good luck hunting down issues!
Another issue is that issues in Ant.Design are mostly filed and debated on GitHub in Chinese. This could be a deal breaker for enterprise applications, but I’m not sure it should be one for early startups since Ant can be used quite minimally, without making use of smarter features like built-in form validation. Still, if you find an issue or bug with the library, it will be difficult to research previous solutions to your issue, and that’s why I recommend making minimal use of the surrounding ecosystem at this stage.
Popular UI libraries for React include Material-UI, Semantic-UI, Foundation, and Bootstrap (this and this), and they are all fairly mature. Material-UI should be singled out as it massively eclipses the others in popularity, with over 22k stargazers — and over 600 open issues. But it turns out that Ant.Design is a surprisingly worthy candidate as well. It’s battle tested by some of the most well-trodden sites on the web (Alibaba, Baidu), and it boasts a brilliant style guide, custom tooling, and, of course, a comprehensive catalogue of components. It also has only 85 open issues at the time of writing, which is a good thing considering its popularity.
So let’s take a tour of the library, see what it has to offer, and how to get started using it.
The Ant components list is dizzying. Sure, it contains the basics — modals, forms (inline and vertical), navigation menus, a grid system. But it also contains a ton of extras, such as a @mentioning system, a timeline, badges, a seriously nice table system, and other small fancy features, such as an involved address box (see the Habitual Residence field). Have a look — it has everything that a modern web application should, with a tasteful, minimalist aesthetic.
There’s a nice, concise section in the documentation on the guiding principles of Ant.Design. I found it a great read as it got me thinking a lot about UI/UX considerations, especially the “Provide an Invitation” section, where they discuss different ways of making interactions discoverable by a user. By the way, if anyone can recommend me a good book on UX, I would be grateful.
The Ant layout system is comprised of a 24-aliquot (a great new word that I learned from the translated documentation — it means parts of a whole) grid and a separate Layout component than you can choose to use. The grid uses the familiar Row/Col system, but you can also specify a prop called flex which allows you to harness Flexbox properties to define a responsive UI. (See a previous blog post of mine for help grokking the Flex standard.)
Flexbox is now fully supported on just about every browser (with partial support on IE 11 as well as some older mobile browsers), so it should be fine to use. If your customer base is largely Internet Explorer users, which does happen in some industries or countries, you would be wise to abstain from using flex Rows or the Layout component, as Layout is built strictly on Flexbox.
Layout includes components for a Sider, Header, Content, and Footer. Again, these are strictly based on Flexbox, so there’s no choice here — but to be honest I’m not sure what these components give you on top of using the standard Row/Col grid system, aside from a couple extra props you can make use of and possibly some built-in design choices. All in all, it doesn’t seem to me to be hugely useful.
Col elements can be supplied with a span prop to define how many aliquots a column takes up and an offset prop to define an optional offset; Row can take a gutter prop to define space between columns in a row (in pixels, not aliquots).
Here’s a UI example from a side project of mine. It contains one row with two columns:
The code would look something like this:
Ant does not let you down as far as forms are concerned, with options for inline, horizontal, and vertical forms, amazing select boxes, and clear validation messages and icons. In fact, it goes a little overboard here. It allows you to wrap your entire form-rendering component in a higher-order component à la Form.create()(<Component />) to gain access to a built-in validator syntax and custom two-way-binding system (cue audible lip biting). You can then specify standard rules such as ‘required’, or supply custom validator methods. (What are Higher Order Components? Check out this excellent post by James K. Nelson.)
Do you need to use their HOC? Absolutely not, and I’m not sure you should. As I said above, going down that path could expose you to language risk should you encounter bugs and I don’t see why you would want to use a custom two-way binding data system anyway. But you could easily use the HOC and just not use the two-way data binding.
Au Naturel — Plain React Forms
So let’s go over how to use the Ant validation messages without using their higher-order component.
Ant gives us three props that we can supply to each Form.Item component to display validation messages or icons:
- validateStatus — This determines the colour & icon scheme of the validation message (see photo above) — valid options are success, warning, error, and validating.
- help — The validation message to display.
- hasFeedback — This is one of them props that don’t require a value. Just include if you want to display the associated icon, and it defaults to true.
Here’s an example of a simple form element that displays a validation message:
Notice that I used the long-form Form.Item component name. You can make yourself a shortcut for this and any other Ant sub-components as follows:const FormItem = Form.Item;// .. allows you to use:
Form Validation using the Ant Higher-Order Component
Now what if we do want to make use of the Ant Form decorator? It’s fairly straightforward to implement. Create your React component class, and then pass it as an argument to Form.create(). The component can then be exported:class SomeComponent extends React.Component
render() <place_form_here.. />
FancyFormComponent = Form.create()(SomeComponent);export FancyFormComponent as default ; // imported as SomeComponent
Inside your form, decorate your Input fields using the getFieldDecorater method, which exposes a ton of extra props on your component. You can now manipulate form elements directly from the props (eek!).
This example in the documentation gives a thorough demonstration on using the complete higher-order component.
Interactive Components — Message (Alert)
Ant provides a number of other components that give web applications a high degree of interactivity. A great example is alerts — or messages, as they’re called in Ant. Adding an alert is as simple as calling message.success('Great! Item has been saved.') in your component. Message types include success, warning, or error. Just don’t forget to import message (lowercase) from ‘antd’.Minimalism at its Best
As I mentioned above, you can either go all-in on the Ant ecosystem (with its custom Webpack adapter), or just opt for the design framework. I went with the latter and I suspect you might too, not the least because using other parts of the ecosystem could require a working knowledge of Chinese. But I’ll cover both options.
Option 1 — Use the CLI
Ant comes with antd-init, a CLI for generating a complete React application with Ant installed. I do not recommend this route for non-Chinese speakers, but if you want to try it, getting started is easy. Just install the CLI using npm, create a new folder, and run antd-init:npm install antd-init -g; mkdir demo-app; cd $_; antd-init;
You will then be greeted by the following message:antd-init@2 is only for experience antd. If you want to create projects, it’s better to init with dva-cli. dva is a redux and react based application framework. elm concept, support side effects, hmr, dynamic load and so on.`
It’s a rabbit’s hole. Open your new application and you will see that your familiar webpack.config.js file is no longer familiar — the CLI uses ant-tool, a “Build Tool Based on Webpack” that I mentioned above. The documentation is in Chinese, but it appears to set common defaults for Webpack and then allow you to just supply values that you want to override. Here’s what the config file looks like:// Learn more on how to config.
// — https://github.com/ant-tool/atool-build#配置扩展module.exports = function(webpackConfig)
]); return webpackConfig;
The index.js contains a lovely demo page that uses the understated Ant styling.
Option 2 — Use Standard Webpack
This would be my preferred route, but it can be more complicated getting your Webpack settings right at first. The Getting Started page includes some good instructions. First install Ant in your React app:$ npm install antd --save
Ant recommends using their own babel-plugin-import in your .babelrc:"presets": [
"plugins": ["transform-decorators-legacy", ..., ["import", [ libraryName: "antd", style: "css" ]]
Make sure your Webpack includes loaders for .js and .css files, and you should be good to go. To use an Ant component, import it in the module file. E.g.import Row, Col, Icon, Button from 'antd';
There’s no doubt that Ant has a lot to offer as a UI framework, with a formidable catalogue of components and a serious ecosystem around it. It does, however, come with some risk. If you experience an issue with the library, you may be stuck communicating in Chinese. Ultimately I recommend trying it out if you like the minimalist aesthetic, while keeping usage of the peripheral Ant ecosystem to a minimum.
ADRIENNE PORTERFELT: You've probably tried to buy somethingonline or had to fill out some formfor your kid's school, and it's reallyhard on your phone.
It's really hard.
MUSTAFA: There seemsto be a dark art when it comes to nativeapplications, and they're such small details thatyou're actually working on.
[MUSIC PLAYING] Quite often, developers willjust throw on input fields onto a page and not reallythink about the UX that's attributed to that.
So they'll think ofa flow of a form, but they won't necessarilyfeel that individual input, so how a user struggles.
And we know thatautocomplete really helps speed up theuser experience and makes fillingforms quite nice.
What are your feelingsand experiences on autocomplete andAutofill as a thing? ADRIENNE PORTER FELT:From the user experience, you've probably triedto buy something online or had to fill out someform for your kid's school.
And it's reallyhard on your phone.
It's really hard.
It's even hard on desktop.
You don't want to get upand get your credit card out of your purse,which is downstairs.
And these browserfeatures just really improve the userexperience of using a form.
In fact, we find that peoplefill out or submit the form 30% faster if the formworks with Autofill.
So we very much suggest thatweb developers think about this.
You want people to besubmitting your forms.
Right? So if you really want yourforms to be submitted quickly, easily, work with Autofill.
MUSTAFA: And why do you thinkdevelopers don't do that? Is it very difficult to do? I'm assuming it's just a fewattributes you add to inputs.
Right? ADRIENNE PORTER FELT: Yeah.
It's actually not that hard.
So you basically need to setup autocomplete attributes, and you need to makesure that you're not doing any fancy things thatreplace the normal select and input elements withother types of elements.
I think most of these developersjust don't think about it, don't realize,that you just need to put a littlebit of extra effort in to make sure that yourform works well with Autofill and to test it out.
MUSTAFA: Is that oneof the challenges, I suppose, like people makingthese custom UI things, which are not native but justlike divs or whatever and replacing that? Is that where thingsfall down with Autofill? ADRIENNE PORTER FELT:Honestly, there's a few things that can go wrong,but that's one of the big ones.
Yeah, so someone wants thisreally beautiful custom form where the dropdown isall fancy and custom.
And as a result of that,they're using all divs.
And the browser can'tfigure out, oh, this is supposed to be a form.
And then in that case,Autofill isn't going to work.
MUSTAFA: And Isuppose there's a lot of accessibilityissues connected to that as well, it looks like.
But from your pointof view, you've got designers anddevelopers, they want to do something custom,like unique experience.
But then as someone whoworks on the browser, you want to say, now letthe browser do the work.
Do you think there'sa middle point there? How can developers at leasthave a custom experience that's unique to theirproduct, but at the same time without breaking standards? Because this is one of thebiggest challenges on the web.
ADRIENNE PORTER FELT:So there's a lot of things you can do tochange the look of the form on the page while stillusing the select and input elements that HTML provides.
Right? You can customizethem in many ways.
I have to admit, there is onething you can't customize, and it's the lookof the dropdown, like in a select element.
But everything else, the wayit statically looks on a page, you can customize.
And the browser will stillknow that they're fields.
What you see is thebiggest challenge then, for Autofillor implementation, from your point of view? ADRIENNE PORTER FELT:I think it's honestly that developers don't thinkabout it, that people don't think to themselves that theyreally need to be testing their forms in this way.
When you've made it yourself,you've filled it out 100 times.
You tested it yourself.
And you don't thinkabout the fact that a user is goingto be coming to it in a different state of mind.
They are tryingto fill it out as fast as they can on the phone.
So I think developersjust aren't really thinking about the factthat they need to take these extra small steps.
MUSTAFA: So in terms ofbrowser compatibility, the things you're using willbe Chrome-specific stuff? Or is that open source-- not open source, butit's cross-platform.
ADRIENNE PORTER FELT: Yeah,that's a good question.
So specifically autocompleteattributes for Autofill, that's standardized.
All the browsers respect them.
With that being said, thereis one part, turning Autofill off-- that's the autocompleteoff attribute-- is not respected by all browsers.
But if you say,this form should be a credit card, thatwill be respected by all the major browsers.
MUSTAFA: But eachexperience, is there slight quirks per browser? Because obviously, that's goingto be a browser-specific thing.
ADRIENNE PORTER FELT: Some are-- they have veryslightly different UIs.
For example, maybe they'llbe integrated with a keyboard widget versus a dropdown.
But I think they're prettysimilar across browsers.
MUSTAFA: You work on theactual Chrome UI itself.
ADRIENNE PORTER FELT: Yes.
MUSTAFA: So are you actuallybuilding that design and code yourself, or are youworking with UX designers in that process? ADRIENNE PORTER FELT: Sowe have a design team.
And the design team helpsus figure out what those UI elements should look like.
We actually have a bigredesign coming up this year that I think is going tomake those substantially more beautiful and also help clean upthe code, which I know that it won't affect most people becauseit won't look any different.
But from our perspective-- MUSTAFA: It's much cleaner.
ADRIENNE PORTER FELT: Man,the code's so much cleaner.
MUSTAFA: What's the actualprocess of you actually creating UI? Because for me, I do front-enddevelopment and code.
But it's like there seemsto be a dark art when it comes to native applications.
And they're such smalldetails that you're actually working on, which the user maynot notice because it works.
But if it's broken,they will see it.
What is the actualprocess that you go through with your Chromedesigners to actually making the UI or testing it? I'm asking lots ofquestions all at once.
ADRIENNE PORTERFELT: Yeah, it's OK.
So I'll talk about theprocess a little bit.
So usually at the beginning,Product, Eng, and the designer will get togetherand talk about what they hope for from the feature.
Often, the designwill then come up with some conceptualmocks of what they feel the feature could look like.
They'll get feedbackfrom Product.
They'll get feedback fromthe engineers, like can we actually build this, whatare the corner cases we need? And then we'lliteratively get closer to what we actually can ship.
So I work on across-platform team, which means that whatwe build has to ship on all of Chrome's platforms.
People think of Chromeas one platform.
But actually, it'sWindows and Mac-- which previouslyhad different UIs, but we're coming toone single standard-- Android, iOS.
And so we have to havedifferent mocks that relate to the specific platforms.
So some things may bepossible for some subset.
Anyway, the designers get allthis feedback from engineers, like, we can do thishere and not there.
And then we iterativelycome through to red lines, which isour final set of designs.
And that's what we implement.
MUSTAFA: So in termsof do the designers work with the actual W3C? Because you'redesigning something which has to beconsidered cross-platform at the same [? time.
?] So likewhen the payment request API, like I was working withsome of the team there, there seems to be thingswhere you have to really be seeing whateveryone else is doing so that the experiencethat you're creating is not so widely different.
And that can be quitechallenging for the designer and developer because youinstinctively want to make it, I don't know, "better.
" But you don't want to makeit so vastly different, because then you stick out.
ADRIENNE PORTER FELT: That'san interesting question.
The challenge here isthat with specs, we try not to specify whatthe UI has to look like.
We try to talk about what theuser experience should be so that we can have the appropriatecallbacks, et cetera, to build that experience.
But we don't like tostandardize the UI itself, which is a fine balance becauseyou have to have a UI in mind when you're designing the API.
But we try to make itas general as possible so that we can build differentUI experiences on top of it.
MUSTAFA: Are you workingwith the browsers as well at the sametime to do that, or is it you do things independently? Because there's the thing.
It's like if you do it[INAUDIBLE] the browsers, then you may be leddown a path that's not the best for everyone.
That it's, OK, it'slike a compromise.
But at the same time, you don'twant that complete disparity.
ADRIENNE PORTER FELT: Itdepends a lot on the standard, honestly.
Some of them willbe heavily driven by Chrome or some other browser.
They really want this API.
They'll drive it, and thenget a little bit of feedback along the way fromother vendors.
Whereas others,from the beginning, there's severaldifferent browser vendors working together.
So honestly, it differsfrom standard to standard.
MUSTAFA: And we're coming to the10-year anniversary of Chrome.
What do you thinkthe future of say, Autofill, or just workingwith the other browsers? Because it seems like thingsare getting much better.
I was speaking toDarren, and it was like, the implementationof Flexbox was a nightmare becausethe standard kept changing.
But with CSS grid, it'samazing that there's so much cross-collaborationbetween the browsers, which is great for the users and thedevelopers working across this.
What do you think the futureis for Chrome as a platform and, I suppose, Autofillas well as a specific? ADRIENNE PORTER FELT: Sofor Autofill specifically, it's hard to saybecause I don't think the limitation there reallyis the lack of browser vendors cooperating.
I feel like actually, there'sbeen a lot of discussion, for example, aroundpayment request.
There's a lot of collaborationbetween Safari and Chrome.
I think that thereal problem we have is that Autofill dependson developer adoption.
Right? If it's hard for thebrowser to figure out what form's in thepage, we're not going to be able to Autofill it.
And so I think thething we really care about is whetherwe can get developers interested in andusing the tools that we have provided for them totry to improve the Autofill experience.
Do you ever have toremove a feature when you see there's not wide adoption? Because I can seefrom an engineer, you're working on Chrome.
You spent your heart andsoul working on this feature.
And then you know it'sgreat for user experience.
You know from theresearch that Autofill will help transactions,and it's just nicer.
But if adoption doesn'thappen, how do you [INAUDIBLE]?? It's the biggest-- ADRIENNE PORTER FELT: Yeah,that's a really hard one.
I've been througha few deprecations.
They can be really challenging.
It's very hard.
So there are a few waysyou can look at it.
One is how many usersinteract with websites that are using such a feature.
And obviously, thatnumber is large, you don't want to create apain point for a lot of users.
But even if thenumber is very small, it might be that there are afew websites, a few companies, whose entire businessmodel depends on having access to this API.
And so that can make itvery difficult, where OK, even if it's thisreally a niche thing, it still can behard to deprecate.
So I think there have beenlots of ongoing discussions in general about howto make that trade-off.
Some of the onesI've been involved in relate specificallyto security and TLS, where if something is makingthe web as a whole as safe, we may have to breaksome connections in order to deprecate it.
And it can be areally painful thing when you've got old serverson the internet that aren't being upgraded.
And maybe it's only a smallpercentage of overall page loads, but it'sstill frustrating when a user is trying to getto a website and it's broken.
But ultimately, fromChrome's perspective, it's the user's experiencethat's paramount.
Right? ADRIENNE PORTER FELT: Yes.
MUSTAFA: And theirsafety and security.
So it's like HTTPS, you couldprobably explain it better, but there's a cutoff pointwhere if your site is not loaded on HTTPS, you're goingto get a message saying, this isn't secure.
Right? ADRIENNE PORTER FELT: Fora long time in Chrome, we showed HTTP as neutral, HTTPas just plain text and no TLS.
MUSTAFA: Sorry, what's TLS,just for my non-designer-- ADRIENNE PORTER FELT: Oh, TLSis the underlying protocol that makes HTTPS HTTPS.
It's why it's secure.
So there is HTTP,which doesn't have any of the end-to-endsecurity bits on it.
And HTTP websiteswere just shown as a neutral standard thing.
Right? Most websites on the web wereHTTP, but that's changed.
We went from a few yearsago, we were at 25% HTTPS.
And now, it's the opposite wherewe're at more like 75% HTTPS.
So made changes in theUI to back that up.
So now when you go to awebsite that says HTTP, it's going to also say"not secure" next to it so people reallyunderstand what that means.
MUSTAFA: That decision mustbe quite tough, though.
In some respects, you need toforce the developers to say users' security is paramount.
But at the sametime, does it feel like you're breaking the web? ADRIENNE PORTER FELT:It doesn't really feel like we'rebreaking the web.
First of all, Ithink people have seen this a long time coming.
We've been talking aboutit for a long time.
We've rolled it out in phases.
So first, we started showing"not secure" specifically for pages with passwordsand credit card form fields.
And then it was for allform fields and web pages when viewed in incognito.
And now we're rolling itout for all HTTP websites.
And as you can see, becauseHTTPS adoption has really increased, it's onlyimpacting less than a quarter of page loads at this point.
MUSTAFA: So really,we're just talking about protecting the user.
ADRIENNE PORTERFELT: Yeah, and I think users have a right to knowthat their information isn't secure when they'regoing to this website and help them make adecision about whether or not they want to keep using thatservice or go to another one.
MUSTAFA: And do youthink users are quite savvy now to see those things? Or is this part of theeducation for the user as well to say, look,there are certain things on the web which arenot secure that you have to take into consideration.
ADRIENNE PORTERFELT: I'll be honest.
We have literally billions ofactive users, so it's hard to-- MUSTAFA: Make ageneral [INAUDIBLE] ADRIENNE PORTERFELT: --say generally whether people are goingto understand it or not.
We think enoughpeople understand it that they have a reactionand that they can reach out to sites saying, hey, Ireally like this site, but I wish it were secure.
And we see people doing thatas we've been rolling out these warnings.
SPEAKER 1: The way thatcellular networks are set up is that there's alwaysthese fringe areas, and there's alwaysthese breakdowns.
And higher latencyis always there.