Tumgik
carmenkleinundevot · 3 months
Text
3 notes · View notes
carmenkleinundevot · 3 years
Text
Eigenheimförderung mit Riester Rente
Neben privaten Sparverträgen, die zum Aufbau einer privaten Altersvorsorge dienen, kann seit einiger Zeit auch das Eigenheim mittels der Riester Zulagen gefördert werden. Diese Eigenheimförderung durch die Riester Rente ist umgangssprachlich auch unter dem Begriff Wohnriester bekannt. Wie funktioniert nun diese Förderung und Inanspruchnahme der Riester-Zulagen in Zusammenhang mit einer Eigenheimfinanzierung? 
Zunächst einmal erhält man die Zulagen auf „ganz normalem“ Wege dadurch, dass man Kapital in einen Sparvertrag einzahlt, der zur Altersvorsorge dienen kann. Normalerweise dürfte man aus diesem Vertrag heraus vor Erreichen des 60. Lebensjahres keine Verfügungen veranlassen, ansonsten müsste man alle bisher erhaltenen Zulagen zurück zahlen. 
kredit in schwierigen fällen schweiz
Genau an dieser Stelle greift im Grunde das Wohnriester in Form der Eigenheimförderung. Man hat nämlich nun die Möglichkeit, bis zu 100 Prozent des bisher angesparten Kapital aus dem Sparvertrag zu entnehmen, und zwar Zulagen unschädlich.Allerdings ist dieses nur dann möglich, wenn man das Kapital dazu nutzt, ein Immobiliendarlehen ganz oder teilweise zu tilgen. Dieses Immobiliendarlehen muss dazu dienen, eine selbst genutzte Immobilie (Eigenheim) zu finanzieren. Für nicht selbst genutzte Objekte besteht also demnach nicht diese Möglichkeit der Eigenheimförderung in Form der Riester Rente. Während man noch vor zwei Jahren mindestens 10.000 Euro und maximal 50.000 Euro des im Sparvertrag angesparten Kapital nutzen nutzen konnte, um diese Tilgung durchzuführen, so gibt es in dieser Hinsicht heute keine Beschränkungen mehr. Man kann also mitunter auch das gesamte Kapital aus dem Sparvertrag für die Tilgung des Immobilienkredites nutzen. Ein weiterer Vorteil dieser neuen Regelung ist zudem für den Kreditnehmer, dass die Banken das angesammelte Kapital, welches sich auf dem Riester-Sparvertrag befindet, im Rahmen der Eigenheimfinanzierung als Eigenkapital betrachten und sich somit der Fremdfinanzierungsbedarf reduziert und man in der summe Kreditzinsen einsparen kann.
Hier finden Sie Tipps zur Umschuldung Trotz Schufa
0 notes
carmenkleinundevot · 5 years
Text
Announcing the New Moz SEO Essentials Certificate: What It Is & How to Get Certified
Posted by BrianChilds
“Does Moz offer a certification?”
Educating the marketing community about SEO is one of our core values here at Moz. I worked at an agency prior to joining the team back in 2016, and much of what I learned about how to deliver SEO to our clients came from reading the Moz Blog and watching MozCon videos.
In 2016, one of Moz’s entrepreneurial product managers, Rachel Moore, launched a new catalog of SEO coursework called Moz Academy. This initiative enabled our community to learn faster through structured, interactive workshops. Since 2017, the team has taught SEO to almost 2,500 students through our various class offerings (I looked it up prior to writing this. That number made me really proud).
Across all these interactions, one question asked by our students kept surfacing: 
Can I get a certificate for completing this coursework? 
For agencies, the ability to show a certificate of completion is a way to differentiate themselves amongst a crowded market. I knew from my own experience how valuable having “HubSpot Inbound Certified” and “Adwords Certified” on my LinkedIn profile was — they allowed our team to show proficiency to our prospective clients. For our friends working as in-house marketers, showing a certificate of completion is a way of showing that the student made good on the investment they requested from their managers.
I’m proud to announce that Moz has put in a tremendous amount of effort to create a certificate program that meets this consistent customer demand. Today, Moz is launching the SEO Essentials Certificate through our Moz Academy platform. Check it out below:
I'm ready to check it out!
What is an SEO Certificate?
An SEO Certificate from Moz is all about developing familiarization with Moz tools and covering some of the essential types of projects you can use to hit the ground running. Though attendees of the Moz Academy come from a variety of backgrounds, we built the certification coursework with an Agency or freelance SEO in mind. However, I believe this material is valuable for anyone interested in learning SEO.
The certificate is focused on five core competency areas:
Fundamental SEO Concepts (Understand the Fundamentals)
Keyword research (Develop Keyword Strategies)
Page optimization (Apply On-Page Optimization Strategies)
Link building essentials (Build Effective Link Strategies)
Reporting on SEO (Create Efficient Reporting Strategies)
By completing this certificate, attendees should be able to articulate for their stakeholders where SEO fits in a digital strategy, how to find and target search engine results pages (SERPs) based on the competitive landscape, and how to approach delivering basic SEO tactics using the Moz toolset.
With this foundation, you'll be able to jump off into more advanced topics such as technical SEO fixes, local SEO, and how to set up your agency for success.
Check out just what's included in the coursework for the Moz SEO Essentials certificate:
1. Understand the Fundamentals
One challenge we observed in the development of SEO coursework: our users often started delivering organic traffic improvements without having a foundational understanding of where SEO tactics fit into a broader digital strategy. Often people will initiate optimization efforts without first conducting effective keyword research. Or, if keyword research was being done, it wasn’t framed within a repeatable, scalable process.
The Understand the Fundamentals course sets the stage for delivering SEO in a way that can be repeated efficiently. In addition to defining essential terminology used in the following classes, you learn how to organize keyword research in a way that produces insight about competition. This relatively simple framework can radically improve targeting of your SEO activities, especially for large enterprises that may compete in several different markets simultaneously.
2. Develop Keyword Strategies
After introducing a framework for conducting keyword research, the certificate program dives into a step-by-step process for creating large keyword clusters and identifying the keywords that will produce the best results. In the development of this coursework, we recognized that many articles and resources talk about keyword research but don’t define a repeatable, scalable process for actually doing it. So many articles about SEO promote hacks that may work for a particular use case, but lack step-by-step instructions. We developed this course to provide you with a practical process that can scale alongside your work.
You’ll learn the importance of mapping keyword clusters to the typical sales funnel customers follow as they move from exploration of solutions to purchase. I’ve presented this material in workshops to large enterprises and small companies — every marketing team that's used this process found it valuable.
By the end of the class, you’ll be familiar with the most valuable features of Moz Keyword Explorer and how to organize lists to help you identify and target the best keywords for your stakeholders.
3. Apply On-Page Optimization Strategies
For many websites, you can find quick wins simply by optimizing page attributes that target strategic keywords. For as much as search has evolved in recent years, we still operate primarily in a world where the text on a page defines the value of that page. This course provides an overview of those attributes and their relative importance.
You'll have a clear understanding of how to use site crawl and page optimization tools to identify, prioritize, and begin optimizing pages based on the keyword strategy they developed in the previous class. Often one of the challenges our students have discovered is that they moved too quickly into optimization without first having their strategy defined. This class will show how strategy and implementation fit together.
4. Build Effective Link Strategies
Link building is another practice that, as an agency marketer, I found difficult to scale. Many articles describe the importance of how relevant links relate to ranking, or hacks that produced a particular result for a page, but not how to create a repeatable process.
As you'll discover in the class, link building is more about process than tools. You’ll understand how to use Moz Link Explorer to isolate the best domains to target amongst the thousands you might consider. I use this process myself whenever launching new websites and it turns a week-long project into a few hours of work. For any agency, where time is literally money, driving down the cost of link analysis with Moz tools can be a big windfall.
5. Create Efficient Reporting Strategies
Whether you're working at an enterprise brand or providing digital marketing services, reporting on outcomes is a big part of your job. Because of the challenges inherent in reporting on attribution with SEO strategies, it's vital to understand both how to set up your data and some common ways to tie SEO projects to broader digital marketing initiatives. This course provides a framework for reporting on SEO that you can adjust to suit your needs. You’ll learn how to use Google Analytics and Moz tools to create actionable reports you can share with your team and stakeholders.
SEO Essentials Certificate FAQs
Here are some of the common questions our community has asked us about the SEO Essentials Certificate during the development process.
How do I get SEO Certified?
Moz offers the SEO Essentials Certificate program via the Moz Academy platform. When you visit Moz Academy, you will see the SEO Essentials Certificate program listed in the catalog. All you have to do is select the course and proceed through the login process.
How long does the SEO Essentials Certificate take?
The Moz SEO Essentials coursework consists of several hours of online instruction, as well as a few quizzes and a final exam. The coursework is developed to be completed within a week of starting. Some attendees have completed the coursework in two days, but for most folks, it takes about a week.
Will I get an SEO certificate and LinkedIn badge?
Yes! We've developed a way for you to get both a SEO certificate you can print and a LinkedIn badge to show you've completed the program. When you pass the final exam, you'll find links to both of these assets and instructions on how to generate them.
How long is the SEO Essentials Certificate valid?
The Moz SEO Essentials Certificate is valid for one year after registration. When the certificate expires, you'll need to retake the coursework to maintain your certification. We set the expiration at one year because SEO changes a lot! (Seriously — just take a look at the Google Algorithm Change History.) We want to make sure that you have the most up-to-date information when displaying your credentials online and to stakeholders.
Sign me up!
Find yourself with questions not addressed in this post? Drop them in the comments and we'll do our best to get them answered.
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
Give it up for Your MozCon 2019 Community Speakers!
Posted by cheryldraper
High fives and fist bumps for each and every person who took the time to pitch their hearts out for this years’ six community speaker spots — a whopping 130 entries were submitted!
Our selection committee read, watched, and researched, whittling things down to a shortlist of top contenders and then read, watched, and researched some more to determine if a potential speaker and their talk would be a perfect fit for the MozCon stage.
We take lots of things into account during our review, but ultimately there are three main factors that determine our final selections:
Strength of the pitch (e.g., value, relevance to the audience, etc.)
Can the content reasonably be delivered in the time allotted?
Does it fit with overall programming and agenda?
After much deliberation, we’re confident these six community speakers are going to be a great addition to the MozCon Stage.
Grab a seat and see for yourself!
Ready to meet your MozCon Community Speakers?
Areej AbuAli, Head of SEO at Verve Search
Fixing the Indexability Challenge: A Data-Based Framework
How do you turn an unwieldy 2.5 million-URL website into a manageable and indexable site of just 20,000 pages? Areej will share the methodology and takeaways used to restructure a job aggregator site which, like many large websites, had huge problems with indexability and the rules used to direct robot crawl. This talk will tackle tough crawling and indexing issues, diving into the case study with flow charts to explain the full approach and how to implement it.
Christi Olson, Head of Evangelism, Search at Microsoft
What Voice Means for Search Marketers: Top Findings from the 2019 Report
How can search marketers take advantage of the strengths and weaknesses of today's voice assistants? Diving into three scenarios for informational, navigational, and transactional queries, Christi will share how to use language semantics for better content creation and paid targeting, how to optimize existing content to be voice-friendly (including the new voice schema markup!), and what to expect from future algorithm updates as they adapt to assistants that read responses aloud, no screen required. Highlighting takeaways around voice commerce from the report, this talk will ultimately provide a breakdown on how search marketers can begin to adapt their shopping experience for v-commerce.
Emily Triplett Lentz, Content Strategy Lead at Help Scout
How to Audit for Inclusive Content
Digital marketers have a responsibility to learn to spot the biases that frequently find their way into online copy, replacing them with alternatives that lead to stronger, clearer messaging and that cultivate wider, more loyal and enthusiastic audiences. Last year, Help Scout audited several years of content for unintentionally exclusionary language that associated physical disabilities or mental illness with negative-sounding terms, resulting in improved writing clarity and a stronger brand. You'll learn what inclusive content is, how it helps to engage a larger and more loyal audience, how to conduct an audit of potentially problematic language on a site, and how to optimize for inclusive, welcoming language.
Greg Gifford, Vice President of Search at DealerOn
Dark Helmet's Guide to Local Domination with Google Posts and Q&A
Google Posts and Questions & Answers are two incredibly powerful features of Google My Business, yet most people don't even know they exist. Greg will walk through Google Posts in detail, sharing how they work, how to use them, and tips for optimization based on testing with hundreds of clients. He'll also cover the Q&A section of GMB (a terrifying feature that lets anyone in the community speak for your business), share the results of a research project covering hundreds of clients, share some hilarious examples of Q&A run wild, and explain exactly how to use Q&A the right way to win more local business.
Joelle Irvine, Director, Marketing & Growth at Bookmark Content and Communications
Image & Visual Search Optimization Opportunities
With voice, local, and rich results only rising in importance, how do image and visual search fit into the online shopping ecosystem? Using examples from Google Images, Google Lens, and Pinterest Lens, Joelle will show how image optimization can improve overall customer experience and play a key role in discoverability, product evaluation, and purchase decisions for online shoppers, while at the same time accepting that image recognition technology is not yet perfect. Learn actionable tactics around image optimization, including image framing, categorizing, structured data, and indexing to better optimize for visual search.
Marie Haynes, Owner at Marie Haynes Consulting Inc.
Super-Practical Tips for Improving Your Site's E-A-T
Google has admitted that they measure the concept of "Expertise, Authoritativeness, and Trustworthiness" in their algorithms. If your site is categorized under YMYL (Your Money or Your Life), you absolutely must have good E-A-T in order to rank well. In this talk, you'll learn how Google measures E-A-T and what changes you can make both on site and off in order to outrank your competitors. Using real-life examples, Marie will answer what E-A-T is and how Google measures it, what changes you can make on your site to improve how E-A-T is displayed, and what you can do off-site to improve E-A-T.
Be sure to check out the initial agenda here to get a taste of all the MozCon goodness we've got in store for you.
Snag your ticket!
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
Using the SERP to build your keyword list
Posted by TheMozTeam
This post was originally published on the STAT blog.
Keyword lists keeping you up at night? We feel you — and so does every other SEO. There’s a lot that goes into producing a robust keyword list and having one can make the difference between seeing the whole SERP landscape or getting just a glimpse.
Because we care about how much sleep you’re getting (a healthy eight hours, please), we whipped up a useful guide on our favourite way to keyword list-build, and all you need are three SERP features: the “People also ask” box, related searches, and the “People also search for” box.
We’ll explain why you should give these features a test drive and how you can get your hands on all their Google-vetted queries for the ultimate, competition-crushing keyword list.
Watch us turn 3,413 Nikon-related keywords into 25,349 without lifting a pinky finger.
Google-approved search terms 
Each of these features are keyword goldmines — all three of them link to new SERPs from terms that are semantically related to the searcher’s original query. As a result, they provide excellent insight into how users follow-up, narrow down, or refine their searches and reveal relevant topics that may be overlooked.
Google has put a lot of effort (and dollars) into understanding and mapping how topics and queries are linked, and these SERP features are the direct result of all that research — Google is literally pointing you to how and what everyone is searching. Which is why we dig them so much.
The "People Also Ask"
You’re probably quite familiar with this accordion-like feature. The “People also ask” box contains questions related to the searcher’s initial query, which then expand to reveal answers that Google has pulled from other websites.
Not only are PAA questions excellent long-tail additions to your keyword set, they’re also a great resource for content inspiration. The various ways that they express the same basic question can help you expand on topics — one piece of content could easily answer PAA questions such as “What a photographer needs to get started?” and “What tools do I need to be a photographer?”
Just try not to fall down the query rabbit hole. While the PAA box used to surface anywhere from one to four Q&A combos, most are “infinite” now and can easily multiply into the hundreds — giving you a seemingly endless supply of SERPs to track.
Just where are all these questions coming from, though? Are people actually asking them? If you read our previous write up on the PAA, you’ll know that Google is not always selecting these questions based on actual searched queries, as some return zero search volume when tracked.
If that wasn’t enough to raise our eyebrows, errant capitalization or non-capitalization (“how many mm are there in one Metre?”), wonky grammar (“Is aperture and f stop the same thing?”), and odd follow-up question choices (“how do you take a selfie?” for the query [easy to use digital camera]) suggest that many PAA questions are the result of machine learning.
In other words, Google is doing its darndest to understand actual search queries and spit out relevant subsequent searches to save users the effort. And it makes sense for us to be on those SERPs when searchers decide to take them up on the offer.
In order to capture all the goodies hiding in a PAA, we created a handy report. For each of your keywords that return a PAA box, our .CSV report will list the questions “also asked” (don’t worry, you’ll only get the number of PAAs that exist before things get infinitely overwhelming) and the URLs that Google sourced the answers from, plus the order they appear in.
After we ran the report for our Nikon queries, we found ourselves looking at 2,838 potential new keywords. A quick scan revealed that many of our PAA boxes returned the same questions over and over again (65.57 percent were duplicates), so we set about removing those. This narrowed our PAA keyword list down to 977 topically related queries to explore.
Related searches 
Another go-to for keyword inspiration are the eight related searches found at the bottom of the SERP that, when clicked, become the search query of a new SERP.
For instance, if we’re interested in ranking for “best professional cameras,” a quick look at the related searches will reveal alternative SERPs that Google thinks our searchers may be interested in, like “best professional camera for beginner,” “best dslr camera,” and “best point and shoot camera.” They help us understand how our searcher may refine or expand upon their original query.
Our related searches report makes it so that you don’t have to manually gather the “Searches related to” yourself — it takes them all and combines them into a crisp and clean .CSV spreadsheet.
This report surfaced 12,526 keywords for Nikon, and just like with our PAA suggestions, we noticed a bunch of repeat related search offenders. After trimming out the duplicates (55.09 percent), we were left with 5,626 unique keywords to help us flesh out our Nikon project.
The "People Also Search for" box 
The term “People also search for” (PASF) isn’t new to the SERP, the feature did get a major refresh back in February, which levelled things up.
Now, instead of just being attached to a knowledge graph, the PASF box also attaches itself to organic URLs and contains extra queries (up to eight on desktop; six on mobile) related to the URL that surfaces it. It’s Google’s way of saying, “Didn’t find what you’re looking for? We’ve got you — try these instead.”
This SERP feature requires you to do a little pogo-sticking in order to surface it — you need to click on the organic search result and then navigate back to the SERP before it materializes.
Obviously collecting these terms would involve a lot of work and potential finger cramps. Thankfully, there’s a handy hack to bypass all that, which is great if pogo-sticking isn’t your cup of tea. This lovely bit of JavaScript code originated from Carlos Canterello and reveals all the PASF boxes on a SERP without all the back and forth-ing.
Or, for those of you feeling DIY-y, you can pull all the raw HTML SERPs and parse them yourself — sans pogo stick, sans hack. Since we’re card-carrying data nerds, we opted for this route — we pulled the raw HTML SERPs through the STAT API and had ourselves a parsing party.
With upwards of eight PASF terms per organic result per SERP, we had oodles of keyword ideas on hand — a grand total of 59,284 to be exact (woah). Once we took away the duplicates, we were left with 18,746 unique keywords. That’s quite a drop from our original number — a whopping 68.38 percent of our keywords were repeats.
Keyword evaluation
Once our reports finished generating and we’d removed all those duplicates, we had 25,349 brand new keywords from all three features — that’s 642.71 percent more than what we started with.
While we trust Google to offer up excellent suggestions, we want to be sure we’ve got only the most relevant keywords to our project. To do this, we conducted a little keyword audit.
First, we combined all our queries into a master list and did some work to surface what was useful and remove the ones that, straight up, made zero sense, such as: “Russian ammo website,” “wallmart,” and “how to look beautiful in friends marriage,” which is super specific and very odd, but we applaud the level of dedication.
This removed 2,238 keywords from the mix, leaving us with a grand total of 23,111 keywords to creep on.
Satisfied with our brand spanking new list, we loaded those puppies into STAT to follow them around for a couple of days for further vetting.
Since we like it when things are Monica-level organized (and because smart segmentation will be key to making sense of all 23,111 of our keywords), we bagged and tagged our new queries into groups of the SERP features from whence they came so we can track which makes the best suggestions.
With our data hyper-organized, and with our search volume populated, we then selected keywords that returned no search volume and kicked them to the curb. You should do this too if you want to minimize clutter and focus on queries that will drive traffic.
We also decided to remove keywords with a search volume of less than 100. Just remember though: search volume is relative. Decide what constitutes as “low” for you — low search volume may be par for the course for your particular industry or vertical. You may just decide you want to keep low search volume keywords in your toolbox.
The rest is up to you 
Now that you know how to acquire boatloads of relevant keywords straight from Google’s billion-dollar consumer research project (the SERP), it’s time to figure out what your next steps are, which is entirely dependent on your SEO strategy.
Maybe you head straight to optimizing. Perhaps you want to do more vetting, like finding the keywords that surface certain SERP features.
If, for instance, we’re interested in featured snippets and local packs, we’d look to the SERP Features dashboard in STAT to see if any of our new keywords return these features, and then click to get those exact keywords. (We’ve even got a handy dandy write-up on exploring a SERP feature strategy to help get you started.)
Whatever adventure you choose, you’re now armed and ready with a crazy number of keywords, and it’s all thanks to your comprehensive list-building, courtesy of the SERP.
Want to learn how you can get cracking and tracking some more? Reach out to our rad team and request a demo to get your very own personalized walkthrough.
If you’re ready to dig in even deeper, check out how to build an intent-based keyword list to get next-level insight.
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
What Links to Target with Google's Disavow Tool - Whiteboard Friday
Posted by Cyrus-Shepard
Do you need to disavow links in the modern age of Google? Is it safe? If so, which links should you disavow? In this Whiteboard Friday, Cyrus Shepard answers all these questions and more. While he makes it clear that the majority of sites shouldn't have to use Google's Disavow Tool, he provides his personal strategies for those times when using the tool makes sense. How do you decide when to disavow? We'd love to hear your process in the comments below!
Click on the whiteboard image above to open a high-resolution version in a new tab!
Video Transcription
Howdy, Moz fans. Welcome to another edition of Whiteboard Friday. I'm Cyrus Shepard. Today we're going to be talking about a big topic — Google's Disavow Tool. We're going to be discussing when you should use it and what links you should target.
Now, this is kind of a scary topic to a lot of SEOs and webmasters. They're kind of scared of the Disavow Tool. They think, "It's not necessary. It can be dangerous. You shouldn't use it." But it's a real tool. It exists for a reason, and Google maintains it for exactly for webmasters to use it. So today we're going to be covering the scenarios which you might consider using it and what links you should target.
Disclaimer! The vast majority of sites don't need to disavow *anything*
Now I want to start out with a big disclaimer. I want this to be approved by the Google spokespeople. So the big disclaimer is the vast majority of sites don't need to disavow anything. Google has made tremendous progress over the last few years of determining what links to simply ignore. In fact, that was one of the big points of the last Penguin 4.0 algorithm update.
Before Penguin, you had to disavow links all the time. But after Penguin 4.0, Google simply ignored most bad links, emphasis on the word "most." It's not a perfect system. They don't ignore all bad links. We'll come back to that point in a minute. There is a danger in using the Disavow Tool of disavowing good links.
That's the biggest problem I see with people who use the disavow is it's really hard to determine what Google counts as a bad link or a harmful link and what they count as a good link. So a lot of people over-disavow and disavow too many things. So that's something you need to look out for. My final point in the disclaimer is large, healthy sites with good link profiles are more immune to bad links.
So if you are The New York Times or Wikipedia and you have a few spam links pointing to you, it's really not going to hurt you. But if your link profile isn't as healthy, that's something you need to consider. So with those disclaimers out of the way, let's talk about the opposite sort of situations, situations where you're going to want to consider using the Disavow Tool.
Good candidates for using the Disavow Tool
Obviously, if you have a manual penalty. Now, these have decreased significantly since Penguin 4.0. But they still exist. People still get manual penalties. Definitely, that's what the Disavow Tool is for. But there are other situations. 
There was a conversation with Marie Haynes, that was published not too long ago, where she was asking in a Google hangout, "Are there other situations that you can use the disavow other than a penalty, where your links may hurt you algorithmically?"
John Mueller said this certainly was the case, that if you want to disavow those obviously dodgy links that could be hurting you algorithmically, it might help Google trust your link profile a little more. If your link profile isn't that healthy in the first place if you only have a handful of links and some of those are dodgy, you don't have a lot to fall back on.
So disavowing those dodgy links can help Google trust the rest of your link profile a little more. 
1. Penalty examples
Okay, with those caveats out of the way and situations where you do want to disavow, a big question people have is, "Well, what should I disavow?" So I've done this for a number of sites, and these are my standards and I'll share them with you. So good candidates to disavow, the best examples are often what Google will give you when they penalize you.
Again it's a little more rare, but when you do get a link penalty, Google will often provide sample links. They don't tell you all of the links to disavow. But they'll give you sample links, and you can go through and you can look for patterns in your links to see what matches what Google is considering a spammy link. You definitely want to include those in your disavow file. 
2. Link schemes
If you've suffered a drop in traffic, or you think Google is hurting you algorithmically because of your links, obviously if you've participated in link schemes, if you've been a little bit naughty and violated Google's Webmaster Guidelines, you definitely want to take a look at those.
We're talking about links that you paid for or someone else paid for. It's possible someone bought some shady links to try to bring you down, although Google is good at ignoring a lot of those. If you use PBNs. Now I know a lot of black hat SEOs that use PBNs and swear by them. But when they don't work, when you've been hurt algorithmically or you've been penalized or your traffic is down and you're using PBNs, that's a good candidate to put in your disavow file.
3. Non-editorial links
Google has a whole list of non-editorial links. We're going to link to it in the transcript below. But these are links that the webmaster didn't intentionally place, things like widgets, forum spam, signature spam, really shady, dodgy links that you control. A good judge of all of these links is often in the anchor text.
4. $$ Anchor text
Is it a money anchor text? Are these money, high-value keywords? Do you control the anchor text? You can generally tell a really shady link by looking at the anchor text. Is it optimized? Could I potentially benefit? Do I control that?
If the answer is yes to those questions, it's usually a good candidate for the disavow file. 
The "maybe" candidates for using the Disavow Tool
Then there's a whole set of links in a bucket that I call the "maybe" file. You might want to disavow. I oftentimes do, but not necessarily. 
1. Malware
So a lot of these would be malware. You click on a link and it gives you a red browser warning that the site contains spam, or your computer freezes up, those toxic links.
If I were Google, I probably wouldn't want to see those types of links linking to a site. I don't like them linking to me. I would probably throw them in the disavow. 
2. Cloaked sites
These are sites when you click on the link, they show Google one set of results, but a user a different set of results. The way you find these is that when you're searching for your links, it's usually a good idea to look at them using a Googlebot user agent.
If you use Chrome, you can get a browser extension. We'll link to some of these in the post below. But look at everything and see everything through Google's eyes using a Googlebot user agent and you can find those cloaked pages. They're kind of a red flag in terms of link quality. 
3. Shady 404s
Now, what do I mean by a shady 404? You click on the link and the page isn't there, and in fact, maybe the whole domain isn't there. You've got a whole bunch of these. It looks like just something is off about these 404s. The reason I throw these in the disavow file is because usually there's no record of what the link was. It was usually some sort of spammy link.
They were trying to rank for something, and then, for whatever reason, they removed the entire domain or it's removed by the domain registrar. Because I don't know what was there, I usually disavow it. It's not going to help me in the future when Google discovers that it's gone anyway. So it's usually a safe bet to disavow those shady 404s. 
4. Bad neighborhood spam
Finally, sometimes you find those bad neighborhood links in your link profile.
These are things like pills, poker, porn, the three P's of bad neighborhoods. If I were Google and I saw porn linking to my non-porn site, I would consider that pretty shady. Now maybe they'll just ignore it, but I just don't feel comfortable having a lot of these bad, spammy neighborhoods linking to me. So I might consider these to throw in the disavow file as well.
Probably okay — don't necessarily need to disavow
Now finally, we often see a lot of people disavowing links that maybe aren't that bad. Again, I want to go back to the point it's hard to tell what Google considers a good link, a valuable link and a poor link. There is a danger in throwing too much in your disavow file, which a lot of people do. They just throw the whole kitchen sink in there.
If you do that, those links aren't going to count, and your traffic might go down. 
1. Scraper sites
So one thing I don't personally put in my disavow file are scraper sites. You get a good link in an online magazine, and then a hundred other sites copy it. These are scraper sites. Google is picking them up. I don't put those in the disavow file because Google is getting better and better at assigning the authority of those links to the original site. I don't find that putting them in the disavow file has really helped, at least with the sites I work with. 
2. Feeds
The same with feeds. You see a lot of feed links in Google's list in your link report. These are just raw HTML feeds, RSS feeds. Again, for the same reason, unless they are feeds or scraper sites from this list over here. If they are feeds and scrapers of good sites, no need. 
3. Auto-generated spam 
These are sites that are automatically generated by robots and programs. They're usually pretty harmless. Google is pretty good at ignoring them. You can tell the difference between auto-generated spam and link scheme again by the anchor text.
Auto-generated spam usually does not have optimized anchor text. It's usually your page title. It's usually broken. These are really low-quality pages that Google generally ignores, that I would not put in a disavow. 
4. Simple low quality
These are things like directories, pages that you look at and you're like, "Oh, wow, they only have three pages on their site. No one is linking to them."
Leave it up to Google to ignore those, and they generally do a pretty good job. Or Google can count them. For things like this, unless it's obvious, unless you're violating these rules, I like to leave them in. I don't like to include them in the disavow. So we've got our list. 
Pro tips for your disavow file
A few pro tips when you actually put your disavow file together if you choose to do so. 
Disavow domain
If you find one bad link on a spammy domain, it's usually a good idea to disavow the entire domain, because there's a good chance that there are other links on there that you're just not spotting.
So using the domain operator in your disavow file is usually a good idea, unless it's a site like WordPress or something with a lot of subdomains. 
Use Search Console & third-party tools
Where do you find your links to disavow? First choice is generally Search Console, the link report in Search Console, because that's the links that Google is actually using. It is helpful to use third-party tools, such as Moz Link Explorer, Ahrefs, SEMrush, whatever your link index is, and that's because you can sort through the anchor text.
When Google gives you their link report, they don't include the anchor text. It's very helpful to use those anchor text reports, such as you would get in Moz Link Explorer, and you can sort through and you can find your over-optimized anchor text, your spammy anchor text. You can find patterns and sort. That's often really helpful to do that in order to sort your information.
Try removing links
If you have a disavow file, and this happens on a lot of older sites, if you're auditing a site, it's a really good idea to go in and check and see if a disavow file already exists. It's possible it was created prior to Penguin 4.0. It's possible there are a lot of good links in there already, and you can try removing links from that disavow file and see if it helps your rankings, because those older disavow files often contain a lot of links that are actually good, that are actually helping you.
Record everything and treat it as an experiment
Finally, record everything. Treat this as any other SEO process. Record everything. Think of it as an experiment. If you disavow, if you make a mistake and your rankings drop or your rankings go up, you want to know what caused that, and you need to be responsible for that and be a good SEO. All right, that's all we have for today.
Leave your own disavow comments below. If you like this video, please share. Thanks, everybody.
Bonus: I really liked these posts for detailing alternative ways of finding links to disavow, so I thought I'd share: 
Too Many Links: Strategies for Disavow & Cleanup
Google’s “Disavow Links Tool”: The Complete Guide
Video transcription by Speechpad.com
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
Restaurant Local SEO: The Google Characteristics of America’s Top-Ranked Eateries
Posted by MiriamEllis
“A good chef has to be a manager, a businessman and a great cook. To marry all three together is sometimes difficult.” - Wolfgang Puck
I like this quote. It makes me hear phones ringing at your local search marketing agency, with aspiring chefs and restaurateurs on the other end of the line, ready to bring experts aboard in the “sometimes difficult” quest for online visibility.
Is your team ready for these clients? How comfortable do you feel talking restaurant Local SEO when such calls come in? When was the last time you took a broad survey of what’s really ranking in this specialized industry?
Allow me to be your prep cook today, and I’ll dice up “best restaurant” local packs for major cities in all 50 US states. We’ll julienne Google Posts usage, rough chop DA, make chiffonade of reviews, owner responses, categories, and a host of other ingredients to determine which characteristics are shared by establishments winning this most superlative of local search phrases.
The finished dish should make us conversant with what it takes these days to be deemed “best” by diners and by Google, empowering your agency to answer those phones with all the breezy confidence of Julia Child.
Methodology
I looked at the 3 businesses in the local pack for “best restaurants (city)” in a major city in each of the 50 states, examining 11 elements for each entry, yielding 4,950 data points. I set aside the food processor for this one and did everything manually. I wanted to avoid the influence of proximity, so I didn’t search for any city in which I was physically located. The results, then, are what a traveler would see when searching for top restaurants in destination cities.
Restaurant results
Now, let’s look at each of the 11 data points together and see what we learn. Take a seat at the table!
Categories prove no barrier to entry
Which restaurant categories make up the dominant percentage of local pack entries for our search?
You might think that a business trying to rank locally for “best restaurants” would want to choose just “restaurant” as their primary Google category as a close match. Or, you might think that since we’re looking at best restaurants, something like “fine dining restaurants” or the historically popular “French restaurants” might top the charts.
Instead, what we’ve discovered is that restaurants of every category can make it into the top 3. Fifty-one percent of the ranking restaurants hailed from highly diverse categories, including Pacific Northwest Restaurant, Pacific Rim Restaurant, Organic, Southern, Polish, Lebanese, Eclectic and just about every imaginable designation. American Restaurant is winning out in bulk with 26 percent of the take, and an additional 7 percent for New American Restaurant. I find this an interesting commentary on the nation’s present gustatory aesthetic as it may indicate a shift away from what might be deemed fancy fare to familiar, homier plates.
Overall, though, we see the celebrated American “melting pot” perfectly represented when searchers seek the best restaurant in any given city. Your client’s food niche, however specialized, should prove no barrier to entry in the local packs.
High prices don’t automatically equal “best”
Do Google’s picks for “best restaurants” share a pricing structure?
It will cost you more than $1000 per head to dine at Urasawa, the nation’s most expensive eatery, and one study estimates that the average cost of a restaurant meal in the US is $12.75. When we look at the price attribute on Google listings, we find that the designation “best” is most common for establishments with charges that fall somewhere in between the economical and the extravagant.
Fifty-eight percent of the top ranked restaurants for our search have the $$ designation and another 25 percent have the $$$. We don’t know Google’s exact monetary value behind these symbols, but for context, a Taco Bell with its $1–$2 entrees would typically be marked as $, while the fabled French Laundry gets $$$$ with its $400–$500 plates. In our study, the cheapest and the costliest restaurants make up only a small percentage of what gets deemed “best.”
There isn’t much information out there about Google’s pricing designations, but it’s generally believed that they stem at least in part from the attribute questions Google sends to searchers. So, this element of your clients’ listings is likely to be influenced by subjective public sentiment. For instance, Californians’ conceptions of priciness may be quite different from North Dakotans’. Nevertheless, on the national average, mid-priced restaurants are most likely to be deemed “best.”
Of anecdotal interest: The only locale in which all 3 top-ranked restaurants were designated at $$$$ was NYC, while in Trenton, NJ, the #1 spot in the local pack belongs to Rozmaryn, serving Polish cuisine at $ prices. It’s interesting to consider how regional economics may contribute to expectations, and your smartest restaurant clients will carefully study what their local market can bear. Meanwhile, 7 of the 150 restaurants we surveyed had no pricing information at all, indicating that Google’s lack of adequate information about this element doesn’t bar an establishment from ranking.
Less than 5 stars is no reason to despair
Is perfection a prerequisite for “best”?
Negative reviews are the stuff of indigestion for restaurateurs, and I’m sincerely hoping this study will provide some welcome relief. The average star rating of the 150 “best” restaurants we surveyed is 4.5. Read that again: 4.5. And the number of perfect 5-star joints in our study? Exactly zero. Time for your agency to spend a moment doing deep breathing with clients.
The highest rating for any restaurant in our data set is 4.8, and only three establishments rated so highly. The lowest is sitting at 4.1. Every other business falls somewhere in-between. These ratings stem from customer reviews, and the 4.5 average proves that perfection is simply not necessary to be “best.”
Breaking down a single dining spot with 73 reviews, a 4.6 star rating was achieved with fifty-six 5-star reviews, four 4-star reviews, three 3-star reviews, two 2-star reviews, and three 1-star reviews. 23 percent of diners in this small review set had a less-than-ideal experience, but the restaurant is still achieving top rankings. Practically speaking for your clients, the odd night when the pho was gummy and the paella was burnt can be tossed onto the compost heap of forgivable mistakes.
Review counts matter, but differ significantly
How many reviews do the best restaurants have?
It’s folk wisdom that any business looking to win local rankings needs to compete on native Google review counts. I agree with that, but was struck by the great variation in review counts across the nation and within given packs. Consider:
The greatest number of reviews in our study was earned by Hattie B’s Hot Chicken in Nashville, TN, coming in at a whopping 4,537!
Meanwhile, Park Heights Restaurant in Tupelo, MS is managing a 3-pack ranking with just 72 reviews, the lowest in our data set.
35 percent of “best”-ranked restaurants have between 100–499 reviews and another 31 percent have between 500–999 reviews. Taken together that’s 66 percent of contenders having yet to break 1,000 reviews.
A restaurant with less than 100 reviews has only a 1 percent chance of ranking for this type of search.
Anecdotally, I don’t know how much data you would have to analyze to be able to find a truly reliable pattern regarding winning review counts. Consider the city of Dallas, where the #1 spot has 3,365 review, but spots #2 and #3 each have just over 300. Compare that to Tallahassee, where a business with 590 reviews is coming in at #1 above a competitor with twice that many. Everybody ranking in Boise has well over 1,000 reviews, but nobody in Bangor is even breaking into the 200s.
The takeaways from this data point is that the national average review count is 893 for our “best” search, but that there is no average magic threshold you can tell a restaurant client they need to cross to get into the pack. Totals vary so much from city to city that your best plan of action is to study the client’s market and strongly urge full review management without making any promise that hitting 1,000 reviews will ensure them beating out that mysterious competitor who is sweeping up with just 400 pieces of consumer sentiment. Remember, no local ranking factor stands in isolation.
Best restaurants aren’t best at owner responses
How many of America’s top chophouses have replied to reviews in the last 60 days?
With a hat tip to Jason Brown at the Local Search Forum for this example of a memorable owner response to a negative review, I’m sorry to say I have some disappointing news. Only 29 percent of the restaurants ranked best in all 50 states had responded to their reviews in the 60 days leading up to my study. There were tributes of lavish praise, cries for understanding, and seething remarks from diners, but less than one-third of owners appeared to be paying the slightest bit of attention.
On the one hand, this indicates that review responsiveness is not a prerequisite for ranking for our desirable search term, but let’s go a step further. In my view, whatever time restaurant owners may be gaining back via unresponsiveness is utterly offset by what they stand to lose if they make a habit of overlooking complaints. Review neglect has been cited as a possible cause of business closure. As my friends David Mihm and Mike Blumenthal always say:“Your brand is its reviews” and mastering the customer service ecosystem is your surest way to build a restaurant brand that lasts.
For your clients, I would look at any local pack with neglected reviews as representative of a weakness. Algorithmically, your client’s active management of the owner response function could become a strength others lack. But I’ll even go beyond that: Restaurants ignoring how large segments of customer service have moved onto the web are showing a deficit of commitment to the long haul. It’s true that some eateries are famous for thriving despite offhand treatment of patrons, but in the average city, a superior commitment to responsiveness could increase many restaurants’ repeat business, revenue and rankings.
Critic reviews nice but not essential
I’ve always wanted to investigate critic reviews for restaurants, as Google gives them a great deal of screen space in the listings:
How many times were critic reviews cited in the Google listings of America’s best restaurants and how does an establishment earn this type of publicity?
With 57 appearances, Lonely Planet is the leading source of professional reviews for our search term, with Zagat and 10Best making strong showings, too. It’s worth noting that 70/150 businesses I investigated surfaced no critic reviews at all. They’re clearly not a requirement for being considered “best”, but most restaurants will benefit from the press. Unfortunately, there are few options for prompting a professional review. To wit:
Lonely Planet — Founded in 1972, Lonely Planet is a travel guide publisher headquartered in Australia. Critic reviews like this one are written for their website and guidebooks simultaneously. You can submit a business for review consideration via this form, but the company makes no guarantees about inclusion.
Zagat — Founded in 1979, Zagat began as a vehicle for aggregating diner reviews. It was purchased by Google in 2011 and sold off to The Infatuation in 2018. Restaurants can’t request Zagat reviews. Instead, the company conducts its own surveys and selects businesses to be rated and reviewed, like this.
10Best — Owned by USA Today Travel Media Group, 10Best employs local writers/travelers to review restaurants and other destinations. Restaurants cannot request a review.
The Infatuation — Founded in 2009 and headquartered in NY, The Infatuation employs diner-writers to create reviews like this one based on multiple anonymous dining experiences that are then published via their app. The also have a SMS-based restaurant recommendation system. They do not accept request from restaurants hoping to be reviewed.
AFAR — Founded in 2009, AFAR is a travel publication with a website, magazine, and app which publishes reviews like this one. There is no form for requesting a review.
Michelin — Founded as a tire company in 1889 in France, Michelin’s subsidiary ViaMichelin is a digital mapping service that houses the reviews Google is pulling. In my study, Chicago, NYC and San Francisco were the only three cities that yielded Michelin reviews like this one and one article states that only 165 US restaurants have qualified for a coveted star rating. The company offers this guide to dining establishments.
As you can see, the surest way to earn a professional review is to become notable enough on the dining scene to gain the unsolicited notice of a critic. 
Google Posts hardly get a seat at best restaurant tables
How many picks for best restaurants are using the Google Posts microblogging feature?
As it turns out, only a meager 16 percent of America’s “best” restaurants in my survey have made any use of Google Posts. In fact, most of the usage I saw wasn’t even current. I had to click the “view previous posts on Google” link to surface past efforts. This statistic is much worse than what Ben Fisher found when he took a broader look at Google Posts utilization and found that 42 percent of local businesses had at least experimented with the feature at some point.
For whatever reason, the eateries in my study are largely neglecting this influential feature, and this knowledge could encompass a competitive advantage for your restaurant clients.
Do you have a restaurateur who is trying to move up the ranks? There is some evidence that devoting a few minutes a week to this form of microblogging could help them get a leg up on lazier competitors.
Google Posts are a natural match for restaurants because they always have something to tout, some appetizing food shot to share, some new menu item to celebrate. As the local SEO on the job, you should be recommending an embrace of this element for its valuable screen real estate in the Google Business Profile, local finder, and maybe even in local packs.
Waiter, there’s some Q&A in my soup
What is the average number of questions top restaurants are receiving on their Google Business Profiles?
Commander’s Palace in New Orleans is absolutely stealing the show in my survey with 56 questions asked via the Q&A feature of the Google Business Profile. Only four restaurants had zero questions. The average number of questions across the board was eight.
As I began looking at the data, I decided not to re-do this earlier study of mine to find out how many questions were actually receiving responses from owners, because I was winding up with the same story. Time and again, answers were being left up to the public, resulting in consumer relations like these:
Takeaway: As I mentioned in a previous post, Greg Gifford found that 40 percent of his clients’ Google Questions were leads. To leave those leads up to the vagaries of the public, including a variety of wags and jokesters, is to leave money on the table. If a potential guest is asking about dietary restrictions, dress codes, gift cards, average prices, parking availability, or ADA compliance, can your restaurant clients really afford to allow a public “maybe” to be the only answer given?
I’d suggest that a dedication to answering questions promptly could increase bookings, cumulatively build the kind of reputation that builds rankings, and possibly even directly impact rankings as a result of being a signal of activity.
A moderate PA & DA gets you into the game
What is the average Page Authority and Domain Authority of restaurants ranking as “best’?
Looking at both the landing page that Google listings are pointing to and the overall authority of each restaurant’s domain, I found that:
The average PA is 36, with a high of 56 and a low of zero being represented by one restaurant with no website link and one restaurant appearing to have no website at all.
The average DA is 41, with a high of 88, one business lacking a website link while actually having a DA of 56 and another one having no apparent website at all. The lowest linked DA I saw was 6.
PA/DA do not = rankings. Within the 50 local packs I surveyed, 32 of them exhibited the #1 restaurant having a lower DA than the establishments sitting at #2 or #3. In one extreme case, a restaurant with a DA of 7 was outranking a website with a DA of 32, and there were the two businesses with the missing website link or missing website. But, for the most part, knowing the range of PA/DA in a pack you are targeting will help you create a baseline for competing.
While pack DA/PA differs significantly from city to city, the average numbers we’ve discovered shouldn’t be out-of-reach for established businesses. If your client’s restaurant is brand new, it’s going to take some serious work to get up market averages, of course.
Local Search Ranking Factors 2019 found that DA was the 9th most important local pack ranking signal, with PA sitting at factor #20. Once you’ve established a range of DA/PA for a local SERP you are trying to move a client up into, your best bet for making improvements will include improving content so that it earns links and powering up your outreach for local links and linktations.
Google’s Local Finder “web results” show where to focus management
Which websites does Google trust enough to cite as references for restaurants?
As it turns out, that trust is limited to a handful of sources:
As the above pie chart shows:
The restaurant’s website was listed as a reference for 99 percent of the candidates in our survey. More proof that you still need a website in 2019, for the very good reason that it feeds data to Google.
Yelp is highly trusted at 76 percent and TripAdvisor is going strong at 43 percent. Your client is likely already aware of the need to manage their reviews on these two platforms. Be sure you’re also checking them for basic data accuracy.
OpenTable and Facebook are each getting a small slice of Google trust, too.
Not shown in the above chart are 13 restaurants that had a web reference from a one-off source, like the Des Moines Register or Dallas Eater. A few very famous establishments, like Brennan’s in New Orleans, surfaced their Wikipedia page, although they didn’t do so consistently. I noticed Wikipedia pages appearing one day as a reference and then disappearing the next day. I was left wondering why.
For me, the core takeaway from this factor is that if Google is highlighting your client’s listing on a given platform as a trusted web result, your agency should go over those pages with a fine-toothed comb, checking for accuracy, activity, and completeness. These are citations Google is telling you are of vital importance.
A few other random ingredients
As I was undertaking this study, there were a few things I noted down but didn’t formally analyze, so consider this as mixed tapas:
Menu implementation is all over the place. While many restaurants are linking directly to their own website via Google’s offered menu link, some are using other services like Single Platform, and far too many have no menu link at all.
Reservation platforms like Open Table are making a strong showing, but many restaurants are drawing a blank on this Google listing field, too. Many, but far from all, of the restaurants designated “best” feature Google’s “reserve a table” function which stems from partnerships with platforms like Open Table and RESY.
Order links are pointing to multiple sources including DoorDash, Postmates, GrubHub, Seamless, and in some cases, the restaurant’s own website (smart!). But, in many cases, no use is being made of this function.
Photos were present for every single best-ranked restaurant. Their quality varied, but they are clearly a “given” in this industry.
Independently-owned restaurants are the clear winners for my search term. With the notable exception of an Olive Garden branch in Parkersburg, WV, and a Cracker Barrel in Bismarck, ND, the top competitors were either single-location or small multi-location brands. For the most part, neither Google nor the dining public associate large chains with “best”.
Honorable mentions go to Bida Manda Laotian Bar & Grill for what looks like a gorgeous and unusual restaurant ranking #1 in Raleigh, NC and to Kermit’s Outlaw Kitchen of Tupelo, MS for the most memorable name in my data set. You can get a lot of creative inspiration from just spending time with restaurant data.
A final garnish to our understanding of this data
I want to note two things as we near the end of our study:
Local rankings emerge from the dynamic scenario of Google’s opinionated algorithms + public opinion and behavior. Doing Local SEO for restaurants means managing a ton of different ingredients: website SEO, link building, review management, GBP signals, etc. We can’t offer clients a generic “formula” for winning across the board. This study has helped us understand national averages so that we can walk into the restaurant space feeling conversant with the industry. In practice, we’ll need to discover the true competitors in each market to shape our strategy for each unique client. And that brings us to some good news.
As I mentioned at the outset of this survey, I specifically avoided proximity as an influence by searching as a traveler to other destinations would. I investigated one local pack for each major city I “visited”. The glad tidings are that, for many of your restaurant clients, there is going to be more than one chance to rank for a search like “best restaurants (city)”. Unless the eatery is in a very small town, Google is going to whip up a variety of local packs based on the searcher’s location. So, that’s something hopeful to share.
What have we learned about restaurant local SEO?
A brief TL;DR you can share easily with your clients:
While the US shows a predictable leaning towards American restaurants, any category can be a contender. So, be bold!
Mid-priced restaurants are considered “best” to a greater degree than the cheapest or most expensive options. Price for your market.
While you’ll likely need at least 100 native Google reviews to break into these packs, well over half of competitors have yet to break the 1,000 mark.
An average 71 percent of competitors are revealing a glaring weakness by neglecting to respond to reviews - so get in there and start embracing customer service to distinguish your restaurant!
A little over half of your competitors have earned critic reviews. If you don’t yet have any, there’s little you can do to earn them beyond becoming well enough known for anonymous professional reviewers to visit you. In the meantime, don’t sweat it.
About three-quarters of your competitors are completely ignoring Google Posts; gain the advantage by getting active.
Potential guests are asking nearly every competitor questions, and so many restaurants are leaving leads on the table by allowing random people to answer. Embrace fast responses to Q&A to stand out from the crowd.
With few exceptions, devotion to authentic link earning efforts can build up your PA/DA to competitive levels.
Pay attention to any platform Google is citing as a resource to be sure the information published there is a complete and accurate.
The current management of other Google Business Profile features like Menus, Reservations and Ordering paints a veritable smorgasbord of providers and a picture of prevalent neglect. If you need to improve visibility, explore every profile field that Google is giving you.
A question for you: Do you market restaurants? Would you be willing to share a cool local SEO tactic with our community? We’d love to hear about your special sauce in the comments below.
Wishing you bon appétit for working in the restaurant local SEO space, with delicious wins ahead!
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
The One-Hour Guide to SEO: Link Building - Whiteboard Friday
Posted by randfish
The final episode in our six-part One-Hour Guide to SEO series deals with a topic that's a perennial favorite among SEOs: link building. Today, learn why links are important to both SEO and to Google, how Google likely measures the value of links, and a few key ways to begin earning your own.
Click on the whiteboard image above to open a high resolution version in a new tab!
Video Transcription
Howdy, Moz fans, and welcome to another edition of Whiteboard Friday. We are back with our final part in the One-Hour Guide to SEO, and this week talking about why links matter to search engines, how you can earn links, and things to consider when doing link building.
Why are links important to SEO?
So we've discussed sort of how search engines rank pages based on the value they provide to users. We've talked about how they consider keyword use and relevant topics and content on the page. But search engines also have this tool of being able to look at all of the links across the web and how they link to other pages, how they point between pages.

So it turns out that Google had this insight early on that what other people say about you is more important, at least to them, than what you say about yourself. So you may say, "I am the best resource on the web for learning about web marketing." But it turns out Google is not going to believe you unless many other sources, that they also trust, say the same thing. Google's big innovation, back in 1997 and 1998, when Sergey Brin and Larry Page came out with their search engine, Google, was PageRank, this idea that by looking at all the links that point to all the pages on the internet and then sort of doing this recursive process of seeing which are the most important and most linked to pages, they could give each page on the web a weight, an amount of PageRank.
Then those pages that had a lot of PageRank, because many people linked to them or many powerful people linked to them, would then pass more weight on when they linked. That understanding of the web is still in place today. It's still a way that Google thinks about links. They've almost certainly moved on from the very simplistic PageRank formula that came out in the late '90s, but that thinking underlies everything they're doing.
How does Google measure the value of links?
Today, Google measures the value of links in many very sophisticated ways, which I'm not going to try and get into, and they're not public about most of these anyway. But there is a lot of intelligence that we have about how they think about links, including things like more important, more authoritative, more well-linked-to pages are going to pass more weight when they link.
A.) More important, authoritative, well-linked-to pages pass more weight when they link
That's true of both individual URLs, an individual page, and websites, a whole website. So for example, if a page on The New York Times links to yoursite.com, that is almost certainly going to be vastly more powerful and influential in moving your rankings or moving your ability to rank in the future than if randstinysite.info — which I haven't yet registered, but I'll get on that — links to yoursite.com.
This weighting, this understanding of there are powerful and important and authoritative websites, and then there are less powerful and important and authoritative websites, and it tends to be the case that more powerful ones tend to provide more ranking value is why so many SEOs and marketers use metrics like Moz's domain authority or some of the metrics from Moz's competitors out in the software space to try and intuit how powerful, how influential will this link be if this domain points to me.
B.) Diversity of domains, rate of link growth, and editorial nature of links ALL matter
So the different kinds of domains and the rate of link growth and the editorial nature of those links all matter. So, for example, if I get many new links from many new websites that have never linked to me before and they are editorially given, meaning I haven't spammed to place them, I haven't paid to place them, they were granted to me because of interesting things that I did or because those sites wanted to editorially endorse my work or my resources, and I do that over time in greater quantities and at a greater rate of acceleration than my competitors, I am likely to outrank them for the words and phrases related to those topics, assuming that all the other smart SEO things that we've talked about in this One-Hour Guide have also been done.
C.) HTML-readable links that don't have rel="nofollow" and contain relevant anchor text on indexable pages pass link benefit
HTML readable links, meaning as a simple text browser browses the web or a simple bot, like Googlebot, which can be much more complex as we talked about in the technical SEO thing, but not necessarily all the time, those HTML readable links that don't have the rel="nofollow" parameter, which is something that you can append to links to say I don't editorially endorse this, and many, many websites do.
If you post a link to Twitter or to Facebook or to LinkedIn or to YouTube, they're going to carry this rel="nofollow,"saying I, YouTube, don't editorially endorse this website that this random user has uploaded a video about. Okay. Well, it's hard to get a link from YouTube. And it contains relevant anchor text on an indexable page, one that Google can actually browse and see, that is going to provide the maximum link benefit.
So a href="https://yoursite.com" great tool for audience intelligence, that would be the ideal link for my new startup, for example, which is SparkToro, because we do audience intelligence and someone saying we're a tool is perfect. This is a link that Google can read, and it provides this information about what we do.
It says great tool for audience intelligence. Awesome. That is powerful anchor text that will help us rank for those words and phrases. There are loads more. There are things like which pages linked to and which pages linked from. There are spam characteristics and trustworthiness of the sources. Alt attributes, when they're used in image tags, serve as the anchor text for the link, if the image is a link.
There's the relationship, the topical relationship of the linking page and linking site. There's text surrounding the link, which I think some tools out there offer you information about. There's location on the page. All of this stuff is used by Google and hundreds more factors to weight links. The important part for us, when we think about links, is generally speaking if you cover your bases here, it's indexable, carries good anchor text, it's from diverse domains, it's at a good pace, it is editorially given in nature, and it's from important, authoritative, and well linked to sites, you're going to be golden 99% of the time.
Are links still important to Google?
Many folks I think ask wisely, "Are links still that important to Google? It seems like the search engine has grown in its understanding of the web and its capacities." Well, there is some pretty solid evidence that links are still very powerful. I think the two most compelling to me are, one, the correlation of link metrics over time. 
So like Google, Moz itself produces an index of the web. It is billions and billions of pages. I think it's actually trillions of pages, trillions of links across hundreds of billions of pages. Moz produces metrics like number of linking root domains to any given domain on the web or any given page on the web.
Moz has a metric called Domain Authority or DA, which sort of tries to best replicate or best correlate to Google's own rankings. So metrics like these, over time, have been shockingly stable. If it were the case someday that Google demoted the value of links in their ranking systems, basically said links are not worth that much, you would expect to see a rapid drop.
But from 2007 to 2019, we've never really seen that. It's fluctuated. Mostly it fluctuates based on the size of the link index. So for many years Ahrefs and Majestic were bigger link indices than Moz. They had better link data, and their metrics were better correlated.
Now Moz, since 2018, is much bigger and has higher correlation than they do. So the various tools are sort of warring with each other, trying to get better and better for their customers. You can see those correlations with Google pretty high, pretty standard, especially for a system that supposedly contains hundreds, if not thousands of elements.
When you see a correlation of 0.25 or 0.3 with one number, linking root domains or page authority or something like that, that's pretty surprising. The second one is that many SEOs will observe this, and I think this is why so many SEO firms and companies pitch their clients this way, which is the number of new, high quality, editorially given linking root domains, linking domains, so The New York Times linked to me, and now The Washington Post linked to me and now wired.com linked to me, these high-quality, different domains, that correlates very nicely with ranking positions.
So if you are ranking number 12 for a keyword phrase and suddenly that page generates many new links from high-quality sources, you can expect to see rapid movement up toward page one, position one, two, or three, and this is very frequent.
How do I get links?
Obviously, this is not alone, but very common. So I think the next reasonable question to ask is, "Okay, Rand, you've convinced me. Links are important. How do I get some?" Glad you asked. There are an infinite number of ways to earn new links, and I will not be able to represent them here. But professional SEOs and professional web marketers often use tactics that fall under a few buckets, and this is certainly not an exhaustive list, but can give you some starting points.
1. Content & outreach
The first one is content and outreach. Essentially, the marketer finds a resource that they could produce, that is relevant to their business, what they provide for customers, data that they have, interesting insights that they have, and they produce that resource knowing that there are people and publications out there that are likely to want to link to it once it exists.
Then they let those people and publications know. This is essentially how press and PR work. This is how a lot of content building and link outreach work. You produce the content itself, the resource, whatever it is, the tool, the dataset, the report, and then you message the people and publications who are likely to want to cover it or link to it or talk about it. That process is tried-and-true. It has worked very well for many, many marketers. 
2. Link reclamation
Second is link reclamation. So this is essentially the process of saying, "Gosh, there are websites out there that used to link to me, that stopped linking." The link broke. The link points to a 404, a page that no longer loads on my website.
The link was supposed to be a link, but they didn't include the link. They said SparkToro, but they forgot to actually point to the SparkToro website. I should drop them a line. Maybe I'll tweet at them, at the reporter who wrote about it and be like, "Hey, you forgot the link." Those types of link reclamation processes can be very effective as well.
They're often some of the easiest, lowest hanging fruit in the link building world. 
3. Directories, resource pages, groups, events, etc.
Directories, resource pages, groups, events, things that you can join and participate in, both online or online and offline, so long as they have a website, often link to your site. The process is simply joining or submitting or sponsoring or what have you.
Most of the time, for example, when I get invited to speak at an event, they will take my biography, a short, three-sentence blurb, that includes a link to my website and what I do, and they will put it on their site. So pitching to speak at events is a way to get included in these groups. I started Moz with my mom, Gillian Muessig, and Moz has forever been a woman-owned business, and so there are women-owned business directories.
I don't think we actually did this, but we could easily go, "Hey, you should include Moz as a woman-owned business.We should be part of your directory here in Seattle." Great, that's a group we could absolutely join and get links from. 
4. Competitors' links
So this is basically the practice you almost certainly will need to use tools to do this. There are some free ways to do it.
The simple, free way to do it is to say, "I have competitor 1 brand name and competitor 2 brand name.I'm going to search for the combination of those two in Google, and I'm going to look for places that have written about and linked to both of them and see if I can also replicate the tactics that got them coverage." The slightly more sophisticated way is to go use a tool. Moz's Link Explorer does this.
So do tools from people like Majestic and Ahrefs. I'm not sure if SEMrush does. But basically you can plug in, "Here's me. Here's my competitors. Tell me who links to them and does not link to me." Moz's tool calls this the Link Intersect function. But you don't even need the link intersect function.
You just plug in a competitor's domain and look at here are all the links that point to them, and then you start to replicate their tactics. There are hundreds more and many, many resources on Moz's website and other great websites about SEO out there that talk about many of these tactics, and you can certainly invest in those. Or you could conceivably hire someone who knows what they're doing to go do this for you. Links are still powerful. 
Okay. Thank you so much. I want to say a huge amount of appreciation to Moz and to Tyler, who's behind the camera — he's waving right now, you can't see it, but he looks adorable waving — and to everyone who has helped make this possible, including Cyrus Shepard and Britney Muller and many others.
Hopefully, this one-hour segment on SEO can help you upgrade your skills dramatically. Hopefully, you'll send it to some other folks who might need to upgrade their understanding and their skills around the practice. And I'll see you again next week for another edition of Whiteboard Friday. Take care.
Video transcription by Speechpad.com
In case you missed them:
Check out the other episodes in the series so far:
The One-Hour Guide to SEO, Part 1: SEO Strategy
The One-Hour Guide to SEO, Part 2: Keyword Research
The One-Hour Guide to SEO, Part 3: Searcher Satisfaction
The One-Hour Guide to SEO, Part 4: Keyword Targeting & On-Page Optimization
The One-Hour Guide to SEO, Part 5: Technical SEO
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
4 Unconventional Ways to Become a Better SEO
Posted by meagar8
Let’s get real for a moment: As much as we hear about positive team cultures and healthy work environments in the digital marketing space, many of us encounter workplace scenarios that are far from the ideal. Some of us might even be part of a team where we feel discouraged to share new ideas or alternative solutions because we know it will be shot down without discussion. Even worse, there are some who feel afraid to ask questions or seek help because their workplace culture doesn’t provide a safe place for learning.
These types of situations, and many others like it, are present in far too many work environments. But what if I told you it doesn’t have to be this way? 
Over the last ten years as a team manager at various agencies, I’ve been working hard to foster a work environment where my employees feel empowered to share their thoughts and can safely learn from their mistakes. Through my experiences, I have found a few strategies to combat negative culture and replace it with a culture of vulnerability and creativity.
Below, I offer four simple steps you can follow that will transform your work environment into one that encourages new ideas, allows for feedback and positive change, and ultimately makes you and your team better digital marketers.
Vulnerability leads to creativity
I first learned about the impact of vulnerability after watching a viral TED talk by Dr. Brene Brown. She defined vulnerability as “uncertainty, risk, and emotional exposure.” She also described vulnerability as “the birthplace of love, belonging, joy, courage, empathy, and creativity.” From this, I learned that to create a culture of vulnerability is to create a culture of creativity. And isn’t creativity at the heart of what we SEOs do?
A culture of vulnerability encourages us to take risks, learn from mistakes, share insights, and deliver top results to our clients. In the fast-paced world of digital marketing, we simply cannot achieve top results with the tactics of yesterday. We also can’t sit around and wait for the next Moz Blog or marketing conference, either. Our best course of action is to take risks, make mistakes, learn from those mistakes, and share insights with others. We have to learn from those with more experience than us and share what we know to those with less experience. In other words, we have to be vulnerable.
Below is a list of four ways you can help create a culture of vulnerability. Whether you are a manager or not, you can impact your team’s culture.
1. Get a second pair of eyes on your next project
Are you finishing up an exciting project for your client? Did you just spend hours of research and implementation to optimize the perfect page? Perfect! Now go ask someone to critique it!
As simple as it sounds, this can make a huge difference in fostering a culture of creativity. It’s also extremely difficult to do.
Large or small, every project or task we complete should be the best your team can provide. All too often, however, team members work in silos and complete these projects without asking for or receiving constructive feedback from their teammates before sending it to the client. This leaves our clients and projects only receiving the best one person can provide rather than the best of an entire team.
We all work with diverse team members that carry varying levels of experience and responsibilities. I bet someone on your team will have something to add to your project that you didn’t already think of. Receiving their feedback means every project that you finish or task that you complete is the best your team has to offer your clients.
Keep in mind, though, that asking for constructive feedback is more than just having someone conduct a “standard QA.” In my experience, a “standard QA” means someone barely looked over what you sent and gave you the thumbs up. Having someone look over your work and provide feedback is only helpful when done correctly.
Say you’ve just completed writing and editing content to a page and you’ve mustered up the courage to have someone QA your work. Rather than sending it over, saying “hey can you review this and make sure I did everything right,” instead try to send detailed instructions like this:
"Here is a <LINK> to a page I just edited. Can you take 15 minutes to review it? Specifically, can you review the Title Tag and Description? This is something the client said is important to them and I want to make sure I get it right."
In many cases, you don’t need your manager to organize this for you. You can set this up yourself and it doesn’t have to be a big thing. Before you finish a project or task this week, work with a team member and ask them for help by simply asking them to QA your work. Worried about taking up too much of their time? Offer to swap tasks. Say you’ll QA some of their work if they QA yours.
Insider tip
You will have greater success and consistency if you make QA a mandatory part of your process for larger projects. Any large project like migrating a site to https or conducting a full SEO audit should have a QA process baked into it.
Six months ago I was tasked to present one of our 200+ point site audits to a high profile client. The presentation was already created with over 100 slides of technical fixes and recommendations. I’m normally pretty comfortable presenting to clients, but I was nervous about presenting such technical details to THIS particular client.
Lucky for me, my team already had a process in place for an in-depth QA for projects like this. My six team members got in a room and I presented to them as if they were the client. Yes, that’s right, I ROLE PLAYED! It was unbearably uncomfortable at first. Knowing that each of my team members (who I respect a whole lot) are sitting right in front of me and making notes on every little mistake I make.
After an agonizing 60 minutes of me presenting to my team, I finished and was now ready for the feedback. I just knew the first thing out of their mouths would be something like “do you even know what SEO stands for?” But it wasn’t. Because my team had plenty of practice providing feedback like this in the past, they were respectful and even more so, helpful. They gave me tips on how to better explain canonicalization, helped me alter some visualization, and gave me positive feedback that ultimately left me confident in presenting to the client later that week.
When teams consistently ask and receive feedback, they not only improve their quality of work, but they also create a culture where team members aren’t afraid to ask for help. A culture where someone is afraid to ask for help is a toxic one and can erode team spirit. This will ultimately decrease the overall quality of your team’s work. On the other hand, a culture where team members feel safe to ask for help will only increase the quality of service and make for a safe and fun team working experience.
2. Hold a half-day all hands brainstorm meeting
Building strategies for websites or solving issues can often be the most engaging work that an SEO can do. Yes that’s right, solving issues is fun and I am not ashamed to admit it. As fun as it is to do this by yourself, it can be even more rewarding and infinitely more useful when a team does it together.
Twice a year my team holds a half-day strategy brainstorm meeting. Each analyst brings a client or issues they are struggling to resolve its website performance, client communication, strategy development, etc. During the meeting, each team member has one hour or more to talk about their client/issue and solicit help from the team. Together, the team dives deep into client specifics to help answer questions and solve issues.
Getting the most out of this meeting requires a bit of prep both from the manager and the team.
Here is a high-level overview of what I do.
Before the Meeting
Each Analyst is given a Client/Issue Brief to fill out describing the issue in detail. We have Analysts answer the following 5 questions:
What is the core issue you are trying to solve?
What have you already looked into or tried?
What haven’t you tried that you think might help?
What other context can you provide that will help in solving this issue?
After all client briefs are filled out and about 1-2 days prior to the half day strategy meeting I will share all the completed briefs to the team so they can familiarize themselves with the issues and come prepared to the meeting with ideas.
Day of the Meeting
Each Analyst will have up to an hour to discuss their issue with the team. Afterwards, the team will deep dive into solving it. During the 60 minute span, ideas will be discussed, Analysts will put on their nerd hats and dive deep into Analytics or code to solve issues. All members of the team are working toward a single goal and that is to solve the issue.
Once the issues is solved the Analyst who first outlined the issue will readback the solutions or ideas to solving the issue. It may not take the full 60 minutes to get to a solution. Whether it takes the entire time or not after one issue is solved another team member announces their issue and the team goes at it again.
Helpful tips
Depending on the size of your team, you may need to split up into smaller groups. I recommend 3-5.
You may be tempted to take longer than an hour but in my experience, this doesn’t work. The pressure of solving an issue in a limited amount of time can help spark creativity.
This meeting is one of the most effective ways my team practices vulnerability allowing the creativity flow freely. The structure is such that each team member has a way to provide and receive feedback. My experience has been that each analyst is open to new ideas and earnestly listens to understand the ways they can improve and grow as an analyst. And with this team effort, our clients are benefitting from the collective knowledge of the team rather than a single individual.
3. Solicit characteristic feedback from your team
This step is not for the faint of heart. If you had a hard time asking for someone to QA your work or presenting a site audit in front of your team, then you may find this one to be the toughest to carry out.
Once a year I hold a special meeting with my team. The purpose of the meeting is to provide a safe place where my employees can provide feedback about me with their fellow teammates. In this meeting, the team meets without me and anonymously fills out a worksheet telling me what I should start doing, stop doing, and keep doing.
Why would I subject myself to this, you ask?
How could I not! Being a great SEO is more than just being great at SEO. Wait, what?!? Yes, you read that right. None of us work in silos. We are part of a team, interact with clients, have expectations from bosses, etc. In other words, the work we do isn’t only technical audits or site edits. It also involves how we communicate and interact with those around us.
This special meeting is meant to focus more on our characteristics and behaviors, over our tactics and SEO chops, ensuring that we are well rounded in our skills and open to all types of feedback to improve ourselves.
How to run a keep/stop/start meeting in 4 steps:
Step 1: Have the team meet together for an hour. After giving initial instructions you will leave the room so that it is just your directs together for 45 minutes.
Step 2: The team writes the behaviors they want you to start doing, stop doing, and keep doing. They do this together on a whiteboard or digitally with one person as a scribe.
Step 3: When identifying the behaviors, the team doesn’t need to be unanimous but they do need to mostly agree. Conversely, the team should not just list them all independently and then paste them together to make a long list.
Step 4: After 45 minutes, you re-enter the room and over the next 15 minutes the team tells you about what they have discussed
Here are some helpful tips to keep in mind:
When receiving the feedback from the team you only have two responses you can give, “thank you” or ask a clarifying question.
The feedback needs to be about you and not the business.
Do this more than once. The team will get better at giving feedback over time.
Here is an example of what my team wrote during my first time running this exercise.
Let’s break down why this meeting is so important.
With me not in the room, the team can discuss openly without holding back.
Having team members work together and come to a consensus before writing down a piece of feedback ensures feedback isn’t from a single team member but rather the whole team.
By leaving the team to do it without me, I show as a manager I trust them and value their feedback.
When I come back to the room, I listen and ask for clarification but don’t argue which helps set an example of receiving feedback from others
The best part? I now have feedback that helps me be a better manager. By implementing some of the feedback, I reinforce the idea that I value my team’s feedback and I am willing to change and grow.
This isn’t just for managers. Team members can do this themselves. You can ask your manager to go through this exercise with you, and if you are brave enough, you can have you teammates do this for you as well.
4. Hold a team meeting to discuss what you have learned recently
Up to this point, we have primarily focused on how you can ask for feedback to help grow a culture of creativity. In this final section, we’ll focus more on how you can share what you have learned to help maintain a culture of creativity.
Tell me if this sounds familiar: I show up at work, catch up on industry news, review my client performance, plug away at my to-do list, check on tests I am running and make adjustments, and so on and so forth.
What are we missing in our normal routines? Collaboration. A theme you may have noticed in this post is that we need to work together to produce our best work. What you read in industry news or what you see in client performance should all be shared with team members.
To do this, my team put together a meeting where we can share our findings. Every 2 weeks, my team meets together for an hour and a half to discuss prepared answers to the following four questions.
Question 1: What is something interesting you have read or discovered in the industry?
This could be as simple as sharing a blog post or going more in depth on some research or a test you have done for a client. The purpose is to show that everyone on the team contributes to how we do SEO and helps contribute knowledge to the team.
Question 2: What are you excited about that you are working on right now?
Who doesn’t love geeking out over a fun site audit, or that content analysis that you have been spending weeks to build? This is that moment to share what you love about your job.
Question 3: What are you working to resolve?
Okay, okay, I know. This is the only section in this meeting that talks about issues you might be struggling to solve. But it is so critical!
Question 4: What have you solved?
Brag, brag, brag! Every analyst has an opportunity to share what they have solve. Issues they overcame. How they out-thought Google and beat down the competition.
In conclusion
Creativity is at the heart of what SEOs do. In order to grow in our roles, we need to continue to expand our minds so we can provide stellar performance for our clients. To do this requires us to receive and give out help with others. Only then will we thrive in a culture that allows us to be safely vulnerable and actively creative.
I would love to hear how your team creates a culture of creativity. Comment below your ideas!
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
How Do I Improve My Domain Authority (DA)?
Posted by Dr-Pete
The Short Version: Don't obsess over Domain Authority (DA) for its own sake. Domain Authority shines at comparing your overall authority (your aggregate link equity, for the most part) to other sites and determining where you can compete. Attract real links that drive traffic, and you'll improve both your Domain Authority and your rankings.
Unless you've been living under a rock, over a rock, or really anywhere rock-adjacent, you may know that Moz has recently invested a lot of time, research, and money in a new-and-improved Domain Authority. People who use Domain Authority (DA) naturally want to improve their score, and this is a question that I admit we've avoided at times, because like any metric, DA can be abused if taken out of context or viewed in isolation.
I set out to write a how-to post, but what follows can only be described as a belligerent FAQ ...
Why do you want to increase DA?
This may sound like a strange question coming from an employee of the company that created Domain Authority, but it's the most important question I can ask you. What's your end-goal? Domain Authority is designed to be an indicator of success (more on that in a moment), but it doesn't drive success. DA is not used by Google and will have no direct impact on your rankings. Increasing your DA solely to increase your DA is pointless vanity.
So, I don't want a high DA?
I understand your confusion. If I had to over-simplify Domain Authority, I would say that DA is an indicator of your aggregate link equity. Yes, all else being equal, a high DA is better than a low DA, and it's ok to strive for a higher DA, but high DA itself should not be your end-goal.
So, DA is useless, then?
No, but like any metric, you can't use it recklessly or out of context. Our Domain Authority resource page dives into more detail, but the short answer is that DA is very good at helping you understand your relative competitiveness. Smart SEO isn't about throwing resources at vanity keywords, but about understanding where you realistically have a chance at competing. Knowing that your DA is 48 is useless in a vacuum. Knowing that your DA is 48 and the sites competing on a query you're targeting have DAs from 30-45 can be extremely useful. Likewise, knowing that your would-be competitors have DAs of 80+ could save you a lot of wasted time and money.
But Google says DA isn't real!
This topic is a blog post (or eleven) in and of itself, but I'm going to reduce it to a couple points. First, Google's official statements tend to define terms very narrowly. What Google has said is that they don't use a domain-level authority metric for rankings. Ok, let's take that at face value. Do you believe that a new page on a low-authority domain (let's say DA = 25) has an equal chance of ranking as a high-authority domain (DA = 75)? Of course not, because every domain benefits from its aggregate internal link equity, which is driven by the links to individual pages. Whether you measure that aggregate effect in a single metric or not, it still exists.
Let me ask another question. How do you measure the competitiveness of a new page, that has no Page Authority (or PageRank or whatever metrics Google uses)? This question is a big part of why Domain Authority exists — to help you understand your ability to compete on terms you haven't targeted and for content you haven't even written yet.
Seriously, give me some tips!
I'll assume you've read all of my warnings and taken them seriously. You want to improve your Domain Authority because it's the best authority metric you have, and authority is generally a good thing. There are no magical secrets to improving the factors that drive DA, but here are the main points:
1. Get more high-authority links
Shocking, I know, but that's the long and short of it. Links from high-authority sites and pages still carry significant ranking power, and they drive both Domain Authority and Page Authority. Even if you choose to ignore DA, you know high-authority links are a good thing to have. Getting them is the topic of thousands of posts and more than a couple of full-length novels (well, ok, books — but there's probably a novel and feature film in the works).
2. Get fewer spammy links
Our new DA score does a much better job of discounting bad links, as Google clearly tries to do. Note that "bad" doesn't mean low-authority links. It's perfectly natural to have some links from low-authority domains and pages, and in many cases it's both relevant and useful to searchers. Moz's Spam Score is pretty complex, but as humans we intuitively know when we're chasing low-quality, low-relevance links. Stop doing that.
3. Get more traffic-driving links
Our new DA score also factors in whether links come from legitimate sites with real traffic, because that's a strong signal of usefulness. Whether or not you use DA regularly, you know that attracting links that drive traffic is a good thing that indicates relevance to searches and drives bottom-line results. It's also a good reason to stop chasing every link you can at all costs. What's the point of a link that no one will see, that drives no traffic, and that is likely discounted by both our authority metrics and Google.
You can't fake real authority
Like any metric based on signals outside of our control, it's theoretically possible to manipulate Domain Authority. The question is: why? If you're using DA to sell DA 10 links for $1, DA 20 links for $2, and DA 30 links for $3, please, for the love of all that is holy, stop (and yes, I've seen that almost verbatim in multiple email pitches). If you're buying those links, please spend that money on something more useful, like sandwiches.
Do the work and build the kind of real authority that moves the needle both for Moz metrics and Google. It's harder in the short-term, but the dividends will pay off for years. Use Domain Authority to understand where you can compete today, cost-effectively, and maximize your investments. Don't let it become just another vanity metric.
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
12 Steps to Lightning Page Speed
Posted by WallStreetOasis.com
At Wall Street Oasis, we’ve noticed that every time we focus on improving our page speed, Google sends us more organic traffic. In 2018, our company's website reached over 80 percent of our traffic from organic search. That’s 24.5 million visits. Needless to say, we are very tuned in to how we can continue to improve our user experience and keep Google happy.
We thought this article would be a great way to highlight the specific steps we take to keep our page speed lightning fast and organic traffic healthy. While this article is somewhat technical (page speed is an important and complex subject) we hope it provides website owners and developers with a framework on how to try and improve their page speed.
Quick technical background: Our website is built on top of the Drupal CMS and we are running on a server with a LAMP stack (plus Varnish and memcache). If you are not using MySQL, however, the steps and principles in this article are still relevant for other databases or a reverse proxy.
Ready? Let’s dig in.
5 Steps to speed up the backend
Before we jump into specific steps that can help you speed up your backend, it might help to review what we mean by “backend”. You can think of the backend of everything that goes into storing data, including the database itself and the servers -- basically anything that helps make the website function that you don’t visually interact with. For more information on the difference between the backend vs. frontend, you read this article
Step 1: Make sure you have a Reverse Proxy configured
This is an important first step. For Wall Street Oasis (WSO), we use a reverse proxy called Varnish. It is by far the most critical and fastest layer of cache and serves the majority of the anonymous traffic (visitors logged out). Varnish caches the whole page in memory, so returning it to the visitor is lightning fast.
Step 2: Extend the TTL of that cache
If you have a large database of content (specifically in the 10,000+ URL range) that doesn’t change very frequently, to drive the hit-rate higher on the Varnish caching layer, you can extend the time to live (TTL basically means how long before you flush the object out of the cache).
For WSO, we went all the way up to two weeks (since we were over 300,000 discussions). At any given time, only a few thousand of those forum URLs are active, so it makes sense to heavily cache the other pages. The downside to this is that when you make any sitewide, template or design changes, you have to wait two weeks for it to arrive across all URLs.
Step 3: Warm up the cache
In order to keep our cache “warm," we have a specific process that hits all the URLs in our sitemap. This increases the likelihood of a page being in the cache when a user or Google bot visits those same pages (i.e. our hit rate improves). It also keeps Varnish full of more objects, ready to be accessed quickly.
As you can see from the chart below, the ratio of “cache hits” (green) to total hits (blue+green) is over 93 percent.
Step 4: Tune your database and focus on the slowest queries
On WSO, we use a MySQL database. Make sure you enable the slow queries report and check it at least every quarter. Check the slowest queries using EXPLAIN. Add indexes where needed and rewrite queries that can be optimized.
On WSO, we use a MySQL database. To tune MySQL, you can use the following scripts: https://github.com/major/MySQLTuner-perl and https://github.com/mattiabasone/tuning-primer
Step 5: HTTP headers
Use HTTP2 server push to send resources to the page before they are requested. Just make sure you test which ones should be pushed, first. JavaScript was a good option for us. You can read more about it here.
Here is an example of server push from our Investment Banking Interview Questions URL:
</files/advagg_js/js__rh8tGyQUC6fPazMoP4YI4X0Fze99Pspus1iL4Am3Nr4__k2v047sfief4SoufV5rlyaT9V0CevRW-VsgHZa2KUGc__TDoTqiqOgPXBrBhVJKZ4CapJRLlJ1LTahU_1ivB9XtQ.js>; rel=preload; as=script,</files/advagg_js/js__TLh0q7OGWS6tv88FccFskwgFrZI9p53uJYwc6wv-a3o__kueGth7dEBcGqUVEib_yvaCzx99rTtEVqb1UaLaylA4__TDoTqiqOgPXBrBhVJKZ4CapJRLlJ1LTahU_1ivB9XtQ.js>; rel=preload; as=script,</files/advagg_js/js__sMVR1us69-sSXhuhQWNXRyjueOEy4FQRK7nr6zzAswY__O9Dxl50YCBWD3WksvdK42k5GXABvKifJooNDTlCQgDw__TDoTqiqOgPXBrBhVJKZ4CapJRLlJ1LTahU_1ivB9XtQ.js>; rel=preload; as=script,
Be sure you're using the correct format. If it is a script: <url>; rel=preload; as=script,
If it is a CSS file: <url>; rel=preload; as=style,
7 Steps to speed up the frontend
The following steps are to help speed up your frontend application. The front-end is the part of a website or application that the user directly interacts with. For example, this includes fonts, drop-down menus, buttons, transitions, sliders, forms, etc.
Step 1: Modify the placement of your JavaScript
Modifying the placement of your JavaScript is probably one of the hardest changes because you will need to continually test to make sure it doesn’t break the functionality of your site. 
I’ve noticed that every time I remove JavaScript, I see page speed improve. I suggest removing as much Javascript as you can. You can minify the necessary JavaScript you do need. You can also combine your JavaScript files but use multiple bundles.
Always try to move JavaScript to the bottom of the page or inline. You can also defer or use the async attribute where possible to guarantee you are not rendering blocking. You can read more about moving JavaScript here.
Step 2: Optimize your images
Use WebP for images when possible (Cloudflare, a CDN, does this for you automatically — I’ll touch more on Cloudflare below). It's an image formatting that uses both Lossy compression and lossless compression.
Always use images with the correct size. For example, if you have an image that is displayed in a 2” x 2 ” square on your site, don’t use a large 10” x 10” image. If you have an image that is bigger than is needed, you are transferring more data through the network and the browser has to resize the image for you
Use lazy load to avoid/delay downloading images that are further down the page and not on the visible part of the screen.
Step 3: Optimize your CSS
You want to make sure your CSS is inline. Online tools like this one can help you find the critical CSS to be inlined and will solve the render blocking. Bonus: you'll keep the cache benefit of having separate files.
Make sure to minify your CSS files (we use AdVagg since we are on the Drupal CMS, but there are many options for this depending on your site).  
Try using less CSS. For instance, if you have certain CSS classes that are only used on your homepage, don't include them on other pages. 
Always combine the CSS files but use multiple bundles. You can read more about this step here.
Move your media queries to specific files so the browser doesn't have to load them before rendering the page. For example: <link href="frontpage-sm.css" rel="stylesheet" media="(min-width: 767px)">
If you’d like more info on how to optimize your CSS, check out Patrick Sexton’s interesting post.
Step 4: Lighten your web fonts (they can be HEAVY)
This is where your developers may get in an argument with your designers if you’re not careful. Everyone wants to look at a beautifully designed website, but if you’re not careful about how you bring this design live, it can cause major unintended speed issues. Here are some tips on how to put your fonts on a diet:
Use inline svg for icon fonts (like font awesome). This way you'll reduce the critical chain path and will avoid empty content when the page is first loaded.
Use fontello to generate the font files. This way, you can include only the glyphs you actually use which leads to smaller files and faster page speed.
If you are going to use web fonts, check if you need all the glyphs defined in the font file. If you don’t need Japanese or Arabic characters, for example, see if there is a version with only the characters you need.
Use Unicode range to select the glyphs you need.
Use woff2 when possible as it is already compressed.
This article is a great resource on web font optimization.
Here is the difference we measured when using optimized fonts:
After reducing our font files from 131kb to 41kb and removing one external resource (useproof), the fully loaded time on our test page dropped all the way from 5.1 to 2.8 seconds. That’s a 44 percent improvement and is sure to make Google smile (see below).
Here’s the 44 percent improvement.
Step 5: Move external resources
When possible, move external resources to your server so you can control expire headers (this will instruct the browsers to cache the resource for longer). For example, we moved our Facebook Pixel to our server and cached it for 14 days. This means you’ll be responsible to check updates from time to time, but it can improve your page speed score.
For example, on our Private Equity Interview Questions page it is possible to see how the fbevents.js file is being loaded from our server and the cache control http header is set to 14 days (1209600 seconds)
cache-control: public, max-age=1209600
Step 6: Use a content delivery network (CDN)
What’s a CDN? Click here to learn more.
I recommend using Cloudflare as it makes a lot of tasks much easier and faster than if you were to try and do them on your own server. Here is what we specifically did on Cloudflare's configuration:
Speed
Auto-minify, check all
Under Polish
Enable Brotoli
Enable Mirage
Choose Lossy
Check WebP
Network
Enable HTTP/2 – You can read more about this topic here
No browsers currently support HTTP/2 over an unencrypted connection. For practical purposes, this means that your website must be served over HTTPS to take advantage of HTTP/2. Cloudflare has a free and easy way to enable HTTPS. Check it out here.
Crypto
Under SSL
Choose Flexible
Under TLS 1.3
Choose Enable+0RTT – More about this topic here.
Step 7: Use service workers
Service workers give the site owner and developers some interesting options (like push notifications), but in terms of performance, we’re most excited about how these workers can help us build a smarter caching system.
To learn how to to get service workers up and running on your site, visit this page.
With resources (images, CSS, javascript, fonts, etc) being cached by a service worker, returning visitors will often be served much faster than if there was no worker at all.
Testing, tools, and takeaways
For each change you make to try and improve speed, you can use the following tools to monitor the impact of the change and make sure you are on the right path:
https://www.webpagetest.org
https://developers.google.com/speed/pagespeed/insights
Google Page Speed Insights was updated on November 2018 (Details here). It gives you an incredible number of suggestions on how to improve the page performance for mobile and desktop based on Light House.
We know there is a lot to digest and a lot of resources linked above, but if you are tight on time, you can just start with Step 1 from both the Backend and Front-End sections. These 2 steps alone can make a major difference on their own.
Good luck and let me know if you have any questions in the comments. I’ll make sure João Guilherme, my Head of Technology, is on to answer any questions for the community at least once a day for the first week this is published.
Happy Tuning!
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
The One-Hour Guide to SEO: Technical SEO - Whiteboard Friday
Posted by randfish
We've arrived at one of the meatiest SEO topics in our series: technical SEO. In this fifth part of the One-Hour Guide to SEO, Rand covers essential technical topics from crawlability to internal link structure to subfolders and far more. Watch on for a firmer grasp of technical SEO fundamentals!
Click on the whiteboard image above to open a high resolution version in a new tab!
Video Transcription
Howdy, Moz fans, and welcome back to our special One-Hour Guide to SEO Whiteboard Friday series. This is Part V - Technical SEO. I want to be totally upfront. Technical SEO is a vast and deep discipline like any of the things we've been talking about in this One-Hour Guide.
There is no way in the next 10 minutes that I can give you everything that you'll ever need to know about technical SEO, but we can cover many of the big, important, structural fundamentals. So that's what we're going to tackle today. You will come out of this having at least a good idea of what you need to be thinking about, and then you can go explore more resources from Moz and many other wonderful websites in the SEO world that can help you along these paths.
1. Every page on the website is unique & uniquely valuable
First off, every page on a website should be two things — unique, unique from all the other pages on that website, and uniquely valuable, meaning it provides some value that a user, a searcher would actually desire and want. Sometimes the degree to which it's uniquely valuable may not be enough, and we'll need to do some intelligent things.
So, for example, if we've got a page about X, Y, and Z versus a page that's sort of, "Oh, this is a little bit of a combination of X and Y that you can get through searching and then filtering this way.Oh, here's another copy of that XY, but it's a slightly different version.Here's one with YZ. This is a page that has almost nothing on it, but we sort of need it to exist for this weird reason that has nothing to do, but no one would ever want to find it through search engines."
Okay, when you encounter these types of pages as opposed to these unique and uniquely valuable ones, you want to think about: Should I be canonicalizing those, meaning point this one back to this one for search engine purposes? Maybe YZ just isn't different enough from Z for it to be a separate page in Google's eyes and in searchers' eyes. So I'm going to use something called the rel=canonical tag to point this YZ page back to Z.
Maybe I want to remove these pages. Oh, this is totally non-valuable to anyone. 404 it. Get it out of here. Maybe I want to block bots from accessing this section of our site. Maybe these are search results that make sense if you've performed this query on our site, but they don't make any sense to be indexed in Google. I'll keep Google out of it using the robots.txt file or the meta robots or other things.
2. Pages are accessible to crawlers, load fast, and can be fully parsed in a text-based browser
Secondarily, pages are accessible to crawlers. They should be accessible to crawlers. They should load fast, as fast as you possibly can. There's a ton of resources about optimizing images and optimizing server response times and optimizing first paint and first meaningful paint and all these different things that go into speed.
But speed is good not only because of technical SEO issues, meaning Google can crawl your pages faster, which oftentimes when people speed up the load speed of their pages, they find that Google crawls more from them and crawls them more frequently, which is a wonderful thing, but also because pages that load fast make users happier. When you make users happier, you make it more likely that they will link and amplify and share and come back and keep loading and not click the back button, all these positive things and avoiding all these negative things.
They should be able to be fully parsed in essentially a text browser, meaning that if you have a relatively unsophisticated browser that is not doing a great job of processing JavaScript or post-loading of script events or other types of content, Flash and stuff like that, it should be the case that a spider should be able to visit that page and still see all of the meaningful content in text form that you want to present.
Google still is not processing every image at the I'm going to analyze everything that's in this image and extract out the text from it level, nor are they doing that with video, nor are they doing that with many kinds of JavaScript and other scripts. So I would urge you and I know many other SEOs, notably Barry Adams, a famous SEO who says that JavaScript is evil, which may be taking it a little bit far, but we catch his meaning, that you should be able to load everything into these pages in HTML in text.
3. Thin content, duplicate content, spider traps/infinite loops are eliminated
Thin content and duplicate content — thin content meaning content that doesn't provide meaningfully useful, differentiated value, and duplicate content meaning it's exactly the same as something else — spider traps and infinite loops, like calendaring systems, these should generally speaking be eliminated. If you have those duplicate versions and they exist for some reason, for example maybe you have a printer-friendly version of an article and the regular version of the article and the mobile version of the article, okay, there should probably be some canonicalization going on there, the rel=canonical tag being used to say this is the original version and here's the mobile friendly version and those kinds of things.
If you have search results in the search results, Google generally prefers that you don't do that. If you have slight variations, Google would prefer that you canonicalize those, especially if the filters on them are not meaningfully and usefully different for searchers. 
4. Pages with valuable content are accessible through a shallow, thorough internal links structure
Number four, pages with valuable content on them should be accessible through just a few clicks, in a shallow but thorough internal link structure.
Now this is an idealized version. You're probably rarely going to encounter exactly this. But let's say I'm on my homepage and my homepage has 100 links to unique pages on it. That gets me to 100 pages. One hundred more links per page gets me to 10,000 pages, and 100 more gets me to 1 million.
So that's only three clicks from homepage to one million pages. You might say, "Well, Rand, that's a little bit of a perfect pyramid structure. I agree. Fair enough. Still, three to four clicks to any page on any website of nearly any size, unless we're talking about a site with hundreds of millions of pages or more, should be the general rule. I should be able to follow that through either a sitemap.
If you have a complex structure and you need to use a sitemap, that's fine. Google is fine with you using an HTML page-level sitemap. Or alternatively, you can just have a good link structure internally that gets everyone easily, within a few clicks, to every page on your site. You don't want to have these holes that require, "Oh, yeah, if you wanted to reach that page, you could, but you'd have to go to our blog and then you'd have to click back to result 9, and then you'd have to click to result 18 and then to result 27, and then you can find it."
No, that's not ideal. That's too many clicks to force people to make to get to a page that's just a little ways back in your structure. 
5. Pages should be optimized to display cleanly and clearly on any device, even at slow connection speeds
Five, I think this is obvious, but for many reasons, including the fact that Google considers mobile friendliness in its ranking systems, you want to have a page that loads clearly and cleanly on any device, even at slow connection speeds, optimized for both mobile and desktop, optimized for 4G and also optimized for 2G and no G.
6. Permanent redirects should use the 301 status code, dead pages the 404, temporarily unavailable the 503, and all okay should use the 200 status code
Permanent redirects. So this page was here. Now it's over here. This old content, we've created a new version of it. Okay, old content, what do we do with you? Well, we might leave you there if we think you're valuable, but we may redirect you. If you're redirecting old stuff for any reason, it should generally use the 301 status code.
If you have a dead page, it should use the 404 status code. You could maybe sometimes use 410, permanently removed, as well. Temporarily unavailable, like we're having some downtime this weekend while we do some maintenance, 503 is what you want. Everything is okay, everything is great, that's a 200. All of your pages that have meaningful content on them should have a 200 code.
These status codes, anything else beyond these, and maybe the 410, generally speaking should be avoided. There are some very occasional, rare, edge use cases. But if you find status codes other than these, for example if you're using Moz, which crawls your website and reports all this data to you and does this technical audit every week, if you see status codes other than these, Moz or other software like it, Screaming Frog or Ryte or DeepCrawl or these other kinds, they'll say, "Hey, this looks problematic to us. You should probably do something about this."
7. Use HTTPS (and make your site secure)
When you are building a website that you want to rank in search engines, it is very wise to use a security certificate and to have HTTPS rather than HTTP, the non-secure version. Those should also be canonicalized. There should never be a time when HTTP is the one that is loading preferably. Google also gives a small reward — I'm not even sure it's that small anymore, it might be fairly significant at this point — to pages that use HTTPS or a penalty to those that don't. 
8. One domain > several, subfolders > subdomains, relevant folders > long, hyphenated URLs
In general, well, I don't even want to say in general. It is nearly universal, with a few edge cases — if you're a very advanced SEO, you might be able to ignore a little bit of this — but it is generally the case that you want one domain, not several. Allmystuff.com, not allmyseattlestuff.com, allmyportlandstuff.com, and allmylastuff.com.
Allmystuff.com is preferable for many, many technical reasons and also because the challenge of ranking multiple websites is so significant compared to the challenge of ranking one. 
You want subfolders, not subdomains, meaning I want allmystuff.com/seattle, /la, and /portland, not seattle.allmystuff.com.
Why is this? Google's representatives have sometimes said that it doesn't really matter and I should do whatever is easy for me. I have so many cases over the years, case studies of folks who moved from a subdomain to a subfolder and saw their rankings increase overnight. Credit to Google's reps.
I'm sure they're getting their information from somewhere. But very frankly, in the real world, it just works all the time to put it in a subfolder. I have never seen a problem being in the subfolder versus the subdomain, where there are so many problems and there are so many issues that I would strongly, strongly urge you against it. I think 95% of professional SEOs, who have ever had a case like this, would do likewise.
Relevant folders should be used rather than long, hyphenated URLs. This is one where we agree with Google. Google generally says, hey, if you have allmystuff.com/seattle/ storagefacilities/top10places, that is far better than /seattle- storage-facilities-top-10-places. It's just the case that Google is good at folder structure analysis and organization, and users like it as well and good breadcrumbs come from there.
There's a bunch of benefits. Generally using this folder structure is preferred to very, very long URLs, especially if you have multiple pages in those folders. 
9. Use breadcrumbs wisely on larger/deeper-structured sites
Last, but not least, at least last that we'll talk about in this technical SEO discussion is using breadcrumbs wisely. So breadcrumbs, actually both technical and on-page, it's good for this.
Google generally learns some things from the structure of your website from using breadcrumbs. They also give you this nice benefit in the search results, where they show your URL in this friendly way, especially on mobile, mobile more so than desktop. They'll show home > seattle > storage facilities. Great, looks beautiful. Works nicely for users. It helps Google as well.
So there are plenty more in-depth resources that we can go into on many of these topics and others around technical SEO, but this is a good starting point. From here, we will take you to Part VI, our last one, on link building next week. Take care.
Video transcription by Speechpad.com
In case you missed them:
Check out the other episodes in the series so far:
The One-Hour Guide to SEO, Part 1: SEO Strategy
The One-Hour Guide to SEO, Part 2: Keyword Research
The One-Hour Guide to SEO, Part 3: Searcher Satisfaction
The One-Hour Guide to SEO, Part 4: Keyword Targeting & On-Page Optimization
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
How to Convince Your Boss to Send You to MozCon 2019
Posted by cheryldraper
From networking with your peers to hearing from industry leaders, there are benefits a-plenty to attending conferences. You know that. Your peers know that. But how do you persuade the powers-that-be (aka your boss) that sending you is beneficial for your business? 
To help convince your boss that won't just be lounging pool-side, sipping cocktails on the company dime, we’ve gathered the goods to help you get your boss to greenlight your MozCon attendance.
How to make the case
Business competition is fiercer than ever. What used to make a splash now feels like it’s barely making ripples. Only those who are able to shift tactics with the changing tides of marketing will be able to come out on top.
And that’s exactly what MozCon is going to help you do.
Covering everything a growing marketer needs for a well-balanced marketing diet (SEO, content, strategy, growth), MozCon delivers top-notch talks from hand-selected speakers over three insightful days in July.
There's so much in store for you this year. Here’s just a sampling of what you can expect at this year’s MozCon:
Speakers and content
Our speakers are real practitioners and industry leaders. We work with them to ensure they deliver the best content and insights to the stage to set you up for a year of success. No sales pitches or talking heads here!
Networking
You work hard taking notes, learning new insights, and digesting all of that knowledge — that’s why we think you deserve a little fun in the evenings. It's your chance to decompress with fellow attendees and make new friends in the industry. We host exciting evening networking events that add to the value you'll get from your day of education. Plus, our Birds of a Feather lunch tables allow you to connect with like-minded peers who share similar interests.
High-quality videos to share with your team
About a month or so after the conference, we’ll send you a link to professionally edited videos of every presentation at the conference. Your colleagues won’t get to partake in the morning Top Pot doughnuts or Starbucks coffee (the #FOMO is real), but they will get a chance to learn everything you did, for free.
An on-going supportive group 
Our MozCon Facebook group is incredibly active, and it’s grown to have a life of its own — marketers ask one another SEO questions, post jobs, look for and offer advice and empathy, and more. It’s a great place to find TAGFEE support and camaraderie long after the conference itself has ended.
Great food on site 
We know that conference food isn’t typically worth mentioning, but at MozCon is notorious for its snacking. You can expect two hot meals a day and loads of snacks from local Seattle vendors — in the past we’ve featured a smorgasbord from the likes of Trophy cupcakes, KuKuRuZa popcorn, Starbucks’ Seattle Reserve cold brew.
Swag
No duds here, we do our homework when it comes to selecting swag worthy of keeping. One-of-a-kind Roger Mozbots, a super-soft t-shirt, and more cool stuff you’ll want to take home and show off.
Wear your heart on your sleeve
MozCon and our attendees give back each year through donating Moz dollars towards a charitable organization.
Discounts for subscribers and groups 
Moz Pro subscribers get a whopping $500 off their ticket cost and there are discounts for groups as well, so make sure to take advantage of savings where you can!
Ticket cost
At MozCon our goal is to breakeven, which means we invest all of your ticket prices back into you. Check out the full breakdown of what your MozCon ticket gets you:
But of course, don’t take our word for it! There are some incredible resources available at your fingertips that tout the benefits of attending conferences:
Bizzabo’s 5 reasons to send your staff to industry events
HubSpot's statistics showing the value of meeting face-to-face and networking
Rand's excellent deep-dive into the reasons for and against attending marketing conferences
The ROI of investing in employee education
I'm convinced, now grab my ticket!
Need a little more to get your boss on board? Check out some videos from years past to get a taste for the caliber of our speakers. We’ve also got a call for community speaker pitches (closes at 5 pm PDT on April 15, 2019) so if you’ve been thinking about breaking into the speaking circuit, it could be an amazing opportunity.
Buy ticket, save money, get competitive marketing insights. Everyone wins!
MozCon is one unforgettable experience that lives and grows with you beyond just the three days you spend in Seattle. And there's no time like the present to pitch MozCon to your boss. If they're still stuck on the "why", let them know about our subscriber or group pricing tiers to your boss — you’ll save hundreds of dollars when you do. Just think of all the Keurigs you could get for that communal kitchen! 
Grab your ticket to MozCon!
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
How Bad Was Google's Deindexing Bug?
Posted by Dr-Pete
On Friday, April 5, after many website owners and SEOs reported pages falling out of rankings, Google confirmed a bug that was causing pages to be deindexed:
MozCast showed a multi-day increase in temperatures, including a 105° spike on April 6. While deindexing would naturally cause ranking flux, as pages temporarily fell out of rankings and then reappeared, SERP-monitoring tools aren't designed to separate the different causes of flux.
Can we isolate deindexing flux?
Google's own tools can help us check whether a page is indexed, but doing this at scale is difficult, and once an event has passed, we no longer have good access to historical data. What if we could isolate a set of URLs, though, that we could reasonably expect to be stable over time? Could we use that set to detect unusual patterns? Across the month of February, the MozCast 10K daily tracking set had 149,043 unique URLs ranking on page one. I reduced that to a subset of URLs with the following properties:
They appeared on page one every day in February (28 total times)
The query did not have sitelinks (i.e. no clear dominant intent)
The URL ranked at position #5 or better
Since MozCast only tracks page one, I wanted to reduce noise from a URL "falling off" from, say, position #9 to #11. Using these qualifiers, I was left with a set of 23,237 "stable" URLs. So, how did those URLs perform over time?
Here's the historical data from February 28, 2019 through April 10. This graph is the percentage of the 23,237 stable URLs that appeared in MozCast SERPs:
Since all of the URLs in the set were stable throughout February, we expect 100% of them to appear on February 28 (which the graph bears out). The change over time isn't dramatic, but what we see is a steady drop-off of URLs (a natural occurrence of changing SERPs over time), with a distinct drop on Friday, April 5th, a recovery, and then a similar drop on Sunday, April 7th.
Could you zoom in for us old folks?
Having just switched to multifocal contacts, I feel your pain. Let's zoom that Y-axis a bit (I wanted to show you the unvarnished truth first) and add a trendline. Here's that zoomed-in graph:

The trend-line is in purple. The departure from trend on April 5th and 7th is pretty easy to see in the zoomed-in version. The day-over-day drop on April 5th was 4.0%, followed by a recovery, and then a second, very similar, 4.4% drop.
Note that this metric moved very little during March's algorithm flux, including the March "core" update. We can't prove definitively that the stable URL drop cleanly represents deindexing, but it appears to not be impacted much by typical Google algorithm updates.
What about dominant intent?
I purposely removed queries with expanded sitelinks from the analysis, since those are highly correlated with dominant intent. I hypothesized that dominant intent might mask some of the effects, as Google is highly invested in surfacing specific sites for those queries. Here's the same analysis just for the queries with expanded sitelinks (this yielded a smaller set of 5,064 stable URLs):
Other than minor variations, the pattern for dominant-intent URLs appears to be very similar to the previous analysis. It appears that the impact of deindexing was widespread.
Was it random or systematic?
It's difficult to determine whether this bug was random, affecting all sites somewhat equally, or was systematic in some way. It's possible that restricting our analysis to "stable" URLs is skewing the results. On the other hand, trying to measure the instability of inherently-unstable URLs is a bit nonsensical. I should also note that the MozCast data set is skewed toward so-called "head" terms. It doesn't contain many queries in the very-long tail, including natural-language questions.
One question we can answer is whether large sites were impacted by the bug. The graph below isolates our "Big 3" in MozCast: Wikipedia, Amazon, and Facebook. This reduced us to 2,454 stable URLs. Unfortunately, the deeper we dive, the smaller the data-set gets:
At the same 90–100% zoomed-in scale, you can see that the impact was smaller than across all stable URLs, but there's still a clear pair of April 5th and April 7th dips. It doesn't appear that these mega-sites were immune.
Looking at the day-over-day data from April 4th to 5th, it appears that the losses were widely distributed across many domains. Of domains that had 10-or-more stable URLs on April 4th, roughly half saw some loss of ranking URLs. The only domains that experienced 100% day-over-day loss were those that had 3-or-fewer stable URLs in our data set. It does not appear from our data that deindexing systematically targeted specific sites.
Is this over, and what's next?
As one of my favorite movie quotes says: "There are no happy endings because nothing ever ends." For now, indexing rates appear to have returned to normal, and I suspect that the worst is over, but I can't predict the future. If you suspect your URLs have been deindexed, it's worth manually reindexing in Google Search Console. Note that this is a fairly tedious process, and there are daily limits in place, so focus on critical pages.
The impact of the deindexing bug does appear to be measurable, although we can argue about how "big" 4% is. For something as consequential as sites falling out of Google rankings, 4% is quite a bit, but the long-term impact for most sites should be minimal. For now, there's not much we can do to adapt — Google is telling us that this was a true bug and not a deliberate change.
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
Links as a Google Ranking Factor: A 2019 Study
Posted by EricEnge
Do Links Still Matter?
For the fourth year running, Stone Temple (now a part of Perficient Digital) conducted a study on how much links matter as a ranking factor. We did that using Moz's Link Explorer and in this year's study, we looked at the largest data set yet — 27,000 queries.
Our study used quadratic mean calculations on the Spearman correlations across all 27K tested queries. Not sure what that means? You can learn more about the study methodology here.
The major study components included:
Total number of links to the ranking pages
Moz DA of the links to the ranking pages
Moz PA of the links to the ranking pages
Slicing these calculations into several sub-categories:
Informational vs. commercial queries
Medical vs. Financial vs. Technology vs. All Other queries
We were also able to evaluate just how much the Moz link index had grown for a subset of the queries because we have used the same data on 16K of the 27K queries for three years running (this year's study looked at 9K more queries, but 16K of the queries were in common). In fact, let's start with that data:
That's pretty significant growth! Congrats to Moz on that improvement.
Brief commentary on correlations
Correlation studies attempt to measure whether or not two factors are related to one another in any way. We use correlation studies to help us understand whether or not one factor potentially causes the other It's important to understand that correlation does not prove causation; it simply suggests that it does.
The example I like to share is that there is a strong correlation between the consumption of ice cream and drowning. That does not mean that one causes the other. In fact, the causal factor here is intuitively obvious — hot weather. People eat more ice cream and people do more swimming when it's hot outside.
But, in the case of links, we also have the fact that Google tells us that links still matter. If that's not enough for you, Google still penalizes sites for questionable link-building practices. This is not an area they would invest in unless links matter.
So how do correlation scores work? 
A correlation score scale runs from -1 to 1. A score of 1 means a perfect correlation between two items. So if we have two variables (x and y), whenever x increases in value, so does y. A score of -1 means the exact opposite: whenever x increases in value, y decreases in value. A score of 0 means there is no perceivable relationship whatsoever. When x increases in value, y is equally likely to increase or decrease in value.
Search is a complex environment to evaluate. Google claims to use over 200 ranking factors. Therefore, it's quite unlikely that any one factor will be dominant. High scores are not likely to happen at all and correlation scores of 0.2 or higher already start to suggest (but not prove) the existence of a relationship.
Core study results
Time to dive in! First, let's take a look at the global view across all 27K queries:
This correlation score comes in at a solid 0.293 score. Considering the complexity of the Google algorithm's 200+ ranking factors, having one single factor come in at a correlation score that high indicates a strong level of correlation.
Next, let's take a look at the correlation to Moz DA and Moz PA:
Both DA and PA show strong correlations; in fact, more so than the total number of links to the ranking page.
This is interesting because it does suggest that at some level, the authority of the linking site and the linking page both matter. By the way, in the four years that we've conducted this study, this is the first time that the DA and PA scores have been a stronger indicator of ranking potential than the pure link count.
More broadly, from a link-building strategy perspective, this provides support for the notion that getting links from more authoritative sites is how you should focus that strategy.
Finally, let's take a look at how commercial and informational queries differ:
Now that's interesting — informational queries show a materially higher level of correlation than commercial ones.
From an interpretative perspective, that does not necessarily mean that they matter less. It may just mean that commercial pages get fewer links, so Google has to depend more heavily on other signals. But should those commercial pages happen to draw links for some reason, the impact of the links may still be as high.
Summary
The data still shows a strong correlation between links and rankings. Google's public statements and its actions (in implementing penalties) also tell the same story. In short, links still matter. But we also see a clear indication that the nature and the quality of those links matter too!
Want more information? You can see the Stone Temple link study here. 
Tell us what you think — do links matter as a ranking factor?
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
How to Find Your True Local Competitors
Posted by MiriamEllis
Who are your clients’ true competitors?
It’s a question that’s become harder to answer. What felt like a fairly simple triangulation between Google, brand, and searcher in the early days of the local web has multiplied into a geodesic dome of localization, personalization, intent matching, and other facets.
This evolution from a simple shape to a more complex shape has the local SEO industry starting to understand the need to talk about trends and patterns vs. empirical rankings.
For instance, you might notice that you just can’t deliver client reports that say, “Congratulations, you’re #1” anymore. And that's because the new reality is that there is no #1 for all searchers. A user on the north side of town may see a completely different local pack of results if they go south, or if they modify their search language. An SEO may get a whole different SERP if they search on one rank checking tool vs. another — or even on the same tool, just five minutes later.
Despite all this, you still need to analyze and report — it remains a core task to audit a client’s competitive landscape.
 Today, let’s talk about how we can distill this dynamic, complex environment down to the simplest shapes to understand who your client's true competitors are. I’ll be sharing a spreadsheet to help you and your clients see the trends and patterns that can create the basis for competitive strategy.
Why are competitive audits necessary...and challenging?
Before we dive into a demo, let’s sync up on what the basic point is of auditing local competitors. Essentially, you’re seeking contrast — you stack up two brands side-by-side to discover the metrics that appear to be making one of them dominant in the local or localized organic SERPs.
From there, you can develop a strategy to emulate the successes of the current winner with the goal of meeting and then surpassing them with superior efforts.
But before you start comparing your brand A to their brand B, you’ve got to know who brand B actually is. What obstacles do you face?
1. SERPs are incredibly diversified
A recent STAT whitepaper that looked at 1.2 million keywords says it all: every SERP is a local SERP. And since both local packs and organic results are both subject to the whims of geo-location and geo-modification, incorporating them into your tracking strategy is a must.
To explain, imagine two searchers are sitting on the same couch. One searches for “Mexican restaurant” and the other searches for “Mexican restaurant near me”. Then, they divvy up searching “Mexican restaurant near me” vs. “Mexican restaurant in San Jose”. And, so on. What they see are local packs that are only about 80 percent similar based on Google recognizing different intents. That’s significant variability.
The scenario gets even more interesting when one of the searchers gets up and travels across town to a different zip code. At that point, the two people making identical queries can see local packs that range from only about 26–65 percent similar. In other words, quite different.
Now, let’s say your client wants to rank for seven key phrases — like “Mexican restaurant,” “Mexican restaurant near me,” “Mexican restaurant San Jose,” “best Mexican restaurant,” “cheap Mexican restaurant,” etc. Your client doesn’t have just three businesses to compete against in the local pack; they now have multiple multiples of three!
2) Even good rank tracking tools can be inconsistent
There are many useful local rank tracking tools out there, and one of the most popular comes to us from BrightLocal. I really like the super easy interface of this tool, but there is a consistency issue with this and other tools I’ve tried, which I’ve captured in a screenshot, below.
Here I’m performing the same search at 5-minute intervals, showing how the reported localized organic ranking of a single business vary widely across time.
The business above appears to move from position 5 to position 12. This illustrates the difficulty of answering the question of who is actually the top competitor when using a tool. My understanding is that this type of variability may result from the use of proxies. If you know of a local rank checker that doesn’t do this, please let our community know in the comments.
In the meantime, what I’ve discovered in my own work is that it’s really hard to find a strong and consistent substitute for manually checking which competitors rank where, on the ground. So, let’s try something out together.
The simplest solution for finding true competitors
Your client owns a Mexican restaurant and has seven main keyword phrases they want to compete for. Follow these five easy steps:
Step 1: Give the client a local pack crash course
If the client doesn’t already know, teach them how to perform a search on Google and recognize what a local pack is. Show them how businesses in the pack rank 1, 2, and 3. If they have more questions about local packs, how they show up in results, and how Google ranks content, they can check out our updated Beginners Guide to SEO.
Step 2: Give the client a spreadsheet and a tiny bit of homework
Give the client a copy of this free spreadsheet, filled out with their most desired keyword phrases. Have them conduct seven searches from a computer located at their place of business* and then fill out the spreadsheet with the names of the three competitors they see for each of the seven phrases. Tell them not to pay attention to any of the other fields of the spreadsheet.
*Be sure the client does this task from their business’ physical location as this is the best way to see what searchers in their area will see in the local results. Why are we doing this? Because Google weights proximity of the searcher-to-the-business so heavily, we have to pretend we’re a searcher at or near the business to emulate Google’s “thought process”.
Step 3: Roll up your sleeves for your part of the work
Now it’s your turn. Look up “directions Google” in Google.
Enter your client’s business address and the address of their first competitor. Write down the distance in the spreadsheet. Repeat for every entry in each of the seven local packs. This will take you approximately 10–15 minutes to cover all 21 locations, so make sure you’re doing it on company time to ensure you're on the clock.
Step 4: Get measuring
Now, in the 2nd column of the spreadsheet, note down the greatest distance Google appears to be going to fill out the results for each pack.
Step 5: Identify competitors by strength
Finally, rate the competitors by the number of times each one appears across all seven local packs. Your spreadsheet should now look something like this:
Looking at the example sheet above, we’ve learned that:
Mi Casa and El Juan’s are the dominant competitors in your client’s market, ranking in 4/7 packs. Plaza Azul is also a strong competitor, with a place in 3/7 packs.
Don Pedro’s and Rubio’s are noteworthy with 2/7 pack appearances.
All the others make just one pack appearance, making them basic competitors.
The radius to which Google is willing to expand to find relevant businesses varies significantly, depending on the search term. While they’re having to go just a couple of miles to find competitors for “Mexican restaurant”, they’re forced to go more than 15 miles for a long tail term like “organic Mexican restaurant”.
You now know who the client’s direct competitors are for their most desired searches, and how far Google is willing to go to make up a local pack for each term. You have discovered a pattern of most dominant competition across your client’s top phrases, signaling which players need to be audited to yield clues about which elements are making them so strong.
The pros and cons of the simple search shape
The old song says that it’s a gift to be simple, but there are some drawbacks to my methodology, namely:
You’ll have to depend on the client to help you out for a few minutes, and some clients are not very good at participation, so you’ll need to convince them of the value of their doing the initial searches for you.
Manual work is sometimes tedious.
Scaling this for a multi-location enterprise would be time-consuming.
Some of your clients are going to be located in large cities and will want to know what competitors are showing up for users across town and in different zip codes. Sometimes, it will be possible to compete with these differently-located competitors, but not always. At any rate, our approach doesn’t cover this scenario and you will be stuck with either using tools (with their known inconsistencies), or sending the client across town to search from that locale. This could quickly become a large chore.
Negatives aside, the positives of this very basic exercise are:
Instead of tying yourself to the limited vision of a single local pack and a single set of competitors, you are seeing a trend, a pattern of dominant market-wide competitors.
You will have swiftly arrived at a base set of dominant, strong, and noteworthy competitors to audit, with the above-stated goal of figuring out what’s helping them to win so that you can create a client strategy for emulating and surpassing them.
Your agency will have created a useful view of your client’s market, understanding the difference between businesses that seem very embedded (like Mi Casa) across multiple packs, vs. those (like Taco Bell) that are only one-offs and could possibly be easier to outpace.
You may discover some extremely valuable competitive intel for your client. For example, if Google is having to cast a 15-mile net to find an organic Mexican restaurant, what if your client started offering more organic items on their menu, writing more about this and getting more reviews that mention it? This will give Google a new option, right in town, to consider for local pack inclusion.
It’s really quite fast to do for a single-location business.
Client buy-in should be a snap for any research they’ve personally helped on, and the spreadsheet should be something they can intuitively and immediately understand.
My questions for you
I’d like to close by asking you some questions about your work doing competitive audits for local businesses. I’d be truly interested in your replies as we all work together to navigate the complex shape of Google’s SERPs:
What percentage of your clients “get” that Google’s results have become so dynamic, with different competitors being shown for different queries and different packs being based on searcher location? What percentage of your clients are “there yet” with this concept vs. the old idea of just being #1, period?
I’ve offered you a manual process for getting at trustworthy data on competitors, but as I’ve said, it does take some work. If something could automate this process for you, especially for multi-location clients, would you be interested in hearing more about it?
How often do you do competitive audits for clients? Monthly? Every six months? Annually?
Thanks for responding, and allow me to wish you and your clients a happy and empowering audit!
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes
carmenkleinundevot · 5 years
Text
The One-Hour Guide to SEO: Keyword Targeting & On-Page Optimization - Whiteboard Friday
Posted by randfish
In the fourth part of the One-Hour Guide to SEO, Rand offers up an on-page SEO checklist to start you off on your way towards perfectly optimized and keyword-targeted pages.
If you missed them, check out the other episodes in the series so far:
The One-Hour Guide to SEO, Part 1: SEO Strategy
The One-Hour Guide to SEO, Part 2: Keyword Research
The One-Hour Guide to SEO, Part 3: Searcher Satisfaction
Click on the whiteboard image above to open a high resolution version in a new tab!
Video Transcription
Howdy, Moz fans. Welcome to another edition of our special One-Hour Guide to SEO. We are now on Part IV – Keyword Targeting and On-Page Optimization. So hopefully, you've watched Part III, where we talked about searcher satisfaction, how to make sure searchers are happy with the page content that you create and the user experience that you build for them, as well as Part II, where we talked about keyword research and how to make sure that you are targeting the right words and phrases that searchers are actually looking for, that you think you can actually rank for, and that actually get real organic click-through rate, because Google's zero-click searches are rising.
Now we're into on-page SEO. So this is essentially taking the words and phrases that we know we want to rank for with the content that we know will help searchers accomplish their task. Now how do we make sure that the page is optimal for ranking in Google?
On-page SEO has evolved
Well, this is very different from the way it was years ago. A long time ago, and unfortunately many people still believe this to be true about SEO, it was: How do I stuff my keywords into all the right tags and places on the page? How do I take advantage of things like the meta keywords tag, which hasn't been used in a decade, maybe two? How do I take advantage of putting all the words and phrases stuffed into my title, my URL, my description, my headline, my H2 through H7 tags, all these kinds of things?
Most of that does not matter, but some of it still does. Some of it is still important, and we need to run through what those are so that you give yourself the best possible chance for ranking.
The on-page SEO checklist
So what I've done here is created a sort of brief, on-page SEO checklist. This is not comprehensive, especially on the technical portion, because we're saving that for Part V, the technical SEO section, which we will get into, of this Guide. In this checklist, some of the most important things are on here. 
☑ Descriptive, compelling, keyword-rich title element
Many of the most important things are on here, and those include things like a descriptive, compelling, keyword-rich but not stuffed title element, also called the page title or a title tag. So, for example, if I am a tool website, like toolsource.com — I made that domain name up, I assume it's registered to somebody — and I want to rank for the "best online survey tools," well, "The Best Online Survey Tools for 2019" is a great title tag, and it's very different from best online survey tools, best online survey software, best online survey software 2019. You've seen title tags like that. You've seen pages that contain stuff like that. That is no longer good SEO practices.
So we want that descriptive, compelling, makes me want to click. Remember that this title is also going to show up in the search results as the title of the snippet that your website appears in.
☑ Meta description designed to draw the click
Second, a meta description. This is still used by search engines, not for rankings though. Sort of think of it like ad text. You are drawing a click, or you're attempting to draw the click. So what you want to do is have a description that tells people what's on the page and inspires them, incites them, makes them want to click on your result instead of somebody else's. That's your chance to say, "Here's why we're valuable and useful."
☑ Easy-to-read, sensible, short URL
An easy-to-read, sensible, short URL. For example, toolsource.com/reviews/best-online-surveys-2019. Perfect, very legible, very readable. I see that in the results, I think, "Okay, I know what that page is going to be." I see that copied and pasted somewhere on the web, I think, "I know what's going to be at that URL. That looks relevant to me."
Or reviews.best-online-tools.info. Okay, well, first off, that's a freaking terrible domain name. /oldseqs?ide=17 bunch of weird letters and tab detail equals this, and UTM parameter equals that. I don't know what this is. I don't know what all this means. By the way, having more than one or two URL parameters is very poorly correlated with and not recommended for trying to rank in search results. So you want to try and rewrite these to be more friendly, shorter, more sensible, and readable by a human being. That will help Google as well.
☑ First paragraph optimized for appearing in featured snippets
That first paragraph, the first paragraph of the content or the first few words of the page should be optimized for appearing in what Google calls featured snippets. Now, featured snippets is when I perform a search, for many queries, I don't just see a list of pages. Sometimes I'll see this box, often with an image and a bunch of descriptive text that's drawn from the page, often from the first paragraph or two. So if you want to get that featured snippet, you have to be able to rank on page one, and you need to be optimized to answer the query right in your first paragraph. But this is an opportunity for you to be ranking in position three or four or five, but still have the featured snippet answer above all the other results. Awesome when you can do this in SEO, very, very powerful thing. Featured snippet optimization, there's a bunch of resources on Moz's website that we can point you to there too.
Featured Snippets: From Start to Finish
How to Discover Featured Snippet Opportunities - Whiteboard Friday
How to Target Featured Snippet Opportunities - Whiteboard Friday
☑ Use the keyword target intelligently in...
☑ The headline
So if I'm trying to rank for "best online survey tools," I would try and use that in my headline. Generally speaking, I like to have the headline and the title of the piece nearly the same or exactly the same so that when someone clicks on that title, they get the same headline on the page and they don't get this cognitive dissonance between the two.
☑ The first paragraph
The first paragraph, we talked about. 
☑ The page content
The page's content, you don't want to have a page that's talking about best online survey tools and you never mention online surveys. That would be a little weird. 
☑ Internal link anchors
An internal link anchor. So if other places on your website talk about online survey tools, you should be linking to this page. This is helpful for Google finding it, helpful for visitors finding it, and helpful to say this is the page that is about this on our website.

I do strongly recommend taking the following advice, which is we are no longer in a world where it makes sense to target one keyword per page. For example, best online survey tools, best online survey software, and best online survey tools 2019 are technically three unique keyword phrases. They have different search volumes. Slightly different results will show up for each of them. But it is no longer the case, whereas it was maybe a decade ago, that I would go create a page for each one of those separate things.
Instead, because these all share the same searcher intent, I want to go with one page, just a single URL that targets all the keywords that share the exact same searcher intent. If searchers are looking to find exactly the same thing but with slightly modified or slight variations in how they phrase things, you should have a page that serves all of those keywords with that same searcher intent rather than multiple pages that try to break those up, for a bunch of reasons. One, it's really hard to get links to all those different pages. Getting links just period is very challenging, and you need them to rank.
Second off, the difference between those is going to be very, very subtle, and it will be awkward and seem to Google very awkward that you have these slight variations with almost the same thing. It might even look to them like duplicate or very similar or low-quality content, which can get you down-ranked. So stick to one page per set of shared intent keywords.
☑ Leverage appropriate rich snippet options
Next, you want to leverage appropriate rich snippet options. So, for example, if you are in the recipes space, you can use a schema markup for recipes to show Google that you've got a picture of the recipe and a cooking time and all these different details. Google offers this in a wide variety of places. When you're doing reviews, they offer you the star ratings. Schema.org has a full list of these, and Google's rich snippets markup page offers a bunch more. So we'll point you to both of those as well.
☑ Images on the page employ...
Last, but certainly not least, because image search is such a huge portion of where Google's search traffic comes from and goes to, it is very wise to optimize the images on the page. Image search traffic can now send significant traffic to you, and optimizing for images can sometimes mean that other people will find your images through Google images and then take them, put them on their own website and link back to you, which solves a huge problem. Getting links is very hard. Images is a great way to do it.
☑ Descriptive, keyword-rich filenames
The images on your page should employ descriptive, keyword-rich filenames, meaning if I have one for typeform, I don't want it to be pick one, two or three. I want it to be typeformlogo or typeformsurveysoftware as the name of the file.
☑ Descriptive alt attributes
The alt attribute or alt tag is part of how you describe that for screen readers and other accessibility-focused devices, and Google also uses that text too. 
☑ Caption text (if appropriate)
Caption text, if that's appropriate, if you have like a photograph and a caption describing it, you want to be descriptive of what's actually in the picture.
☑ Stored in same domain and subdomain
These files, in order to perform well, they generally need to be hosted on the same domain and subdomain. If, for example, all your images are stored on an Amazon Web Services domain and you don't bother rewriting or making sure that the domain looks like it's on toolsource.com/photos or /images here, that can cause real ranking problems. Oftentimes you won't perform at all in Google images because they don't associate the image with the same domain. Same subdomain as well is preferable.
If you do all these things and you nail searcher intent and you've got your keyword research, you are ready to move on to technical SEO and link building and then start ranking. So we'll see you for that next edition next week. Take care.
Video transcription by Speechpad.com
Sign up for The Moz Top 10, a semimonthly mailer updating you on the top ten hottest pieces of SEO news, tips, and rad links uncovered by the Moz team. Think of it as your exclusive digest of stuff you don't have time to hunt down but want to read!
0 notes