这厮

observing

  BlogJava :: 首页 :: 联系 :: 聚合  :: 管理
  48 Posts :: 3 Stories :: 3 Comments :: 0 Trackbacks

#

I’m very proud of the business that we’ve created here at Mocality, but I’m especially proud of two things:

  1. Our crowdsourcing program. When we started this business, we knew that (unlike in the UK or US, where you can just kickstart your directory business with a DVD of business data bought from a commercial supplier), if we wanted a comprehensive database of Kenyan business, we would have to build it ourselves. We knew also that if we wanted to build the business quickly, we’d have to engage a lot of Kenyans to help us. So we built our crowd program that utilises M-PESA (Kenya’s ubiquitous Mobile Money system) to reward any Kenyan with a mobile phone who contributes entries to our database, once those entries have been validated by our team. Over two years, we’ve paid out Ksh. 11m (over $100,000) to thousands of individuals, and we have built Kenya’s most comprehensive directory, with over 170,000 verified listings. Personally, I regard the program as one of THE highlights of my 18 year career on the internet.
  2. From day 1, we aimed to target all Kenyan businesses, irrespective of size. As a result, for about 2/3rds of our listed businesses, Mocality is their first step onto the web. That’s about 100,000 businesses that Mocality has brought online.

Please bear these two facts in mind as you read what follows.

Our database IS our business, and we protect and tend it very carefully. We spot and block automated attacks, amongst other measures. We regularly contact our business owners, to help them keep their records up-to-date, and they are welcome to contact our call centre team for help whenever they need it.

In September, Google launched Getting Kenyan Businesses Online (GKBO). Whilst we saw aspects of their program that were competitive, we welcomed the initiative, as Kenya still has enough growth in it that every new entrant helps the overall market. We are also confident enough in our product, our local team, and our deep local commitment that we believe we can hold our own against any competition, playing fair.

Shortly after that launch, we started receiving some odd calls. One or two business owners were clearly getting confused because they wanted help with their website, and we don’t currently offer websites, only a listing. Initially, we didn’t think much of it, but the confusing calls continued through November.

The Forensic analysis

What follows is necessarily a little technical. I’ve tried to make it as clear as I can, but two definitions may help the lay reader:

  • IP Address – the numerical id by which computers identify themselves online.
  • User-Agent - When a browser requests a page from a webserver, it tells the server what make, model, and version of browser it is, so that the webserver can serve content tailored to that browser’s capabilities. Webservers keep a log of both these details for every page requested, allowing us to do interesting detective work.


Conclusion

Since October, Google’s GKBO appears to have been systematically accessing Mocality’s database and attempting to sell their competing product to our business owners. They have been telling untruths about their relationship with us, and about our business practices, in order to do so. As of January 11th, nearly 30% of our database has apparently been contacted.

Furthermore, they now seem to have outsourced this operation from Kenya to India.

When we started this investigation, I thought that we’d catch a rogue call-centre employee, point out to Google that they were violating our Terms and conditions (sections 9.12 and 9.17, amongst others), someone would get a slap on the wrist, and life would continue.

I did not expect to find a human-powered, systematic, months-long, fraudulent (falsely claiming to be collaborating with us, and worse) attempt to undermine our business, being perpetrated from call centres on 2 continents.

Google is a key part of our business strategy. Mocality will succeed if our member businesses are discoverable by people via Google. We actually track how well our businesses place on Google as a key metric, and have always regarded it as a symbiotic relationship. We are in the business of creating local Kenyan content that Google can sell their adwords against. More than 50% of our non-direct traffic comes via Google (paid or organic). For us, the cost of going elsewhere is NOT zero.

Furthermore, we spend a very significant sum on advertising with Google Kenya. I wouldn’t be surprised if we are one of their largest local customers, between Mocality and our sister site Dealfish.co.ke.

Kenya has a comparatively well-educated but poor population and high levels of unemployment. Mocality designed our crowd sourcing program to provide an opportunity for large numbers of people to help themselves by helping us. By apparently systematically trawling our database, and then outsourcing that trawl to another continent, Google isn’t just scalping us, they’re also scalping every Kenyan who has participated in our program.

I moved to Africa from the UK 30 months ago to be CEO of Mocality. When I moved, Kenya’s reputation as a corrupt place to do business made me nervous. I’ve been very happily surprised- until this point, I’ve not done business with any company here that was not completely honestly conducted. It is important for global businesses to adapt to local cultural practice, but ethics are an invariant. As a admirer of Google’s usually bold ethical stance around the world, to find those principles are not applied in Kenya is simply… saddening.

Someone, somewhere, has some questions to answer.

These are my personal top 3:

  • If Google wanted to work with our data, why didn’t they just ask?
    In discussions with various Google Kenya/Africa folks in the past, I’d raised the idea of working together more closely in Kenya. Getting Kenyan businesses online is precisely what we do.
  • Who authorised this? Until we uncovered the ‘India by way of Mountain View’ angle, I could have believed that this was a local team that somehow forgot the corporate motto, but not now.
  • Who knew, and who SHOULD have known, even if they didn’t know?

read full here:

 

posted @ 2012-01-14 21:24 cnbarry 阅读(209) | 评论 (0)编辑 收藏

Search Engine Optimization is growing up. I am not ready to say the Wild West SEO days are completely eradicated, but in 2011 good search engine optimization is less about trickery and more about engaging content and audience development than ever before.

Over the years, quality optimizers have become more prone to avoid technical tricks like using CSS image replacement to inject keyword text or controlling the flow of PageRank by hiding links from search engines.

Search engines keep getting better at crawling and indexing. If you are unwilling to burn your website or risk your career, you follow the search engines’ terms of service.

During 2011 the conservative attitude toward code crossed chasm to apply to content. For years, websites churned-out poorly written, generic articles in the name of long-tail keyword optimization. It worked so well some people turned crappy content into startups.

Now, thanks to Panda, Google’s site-wide penalty for having too much low quality content, people are asking why anyone would put pages on a website that no one wants to read, share or link to? Without taking potshots at the past, most of those articles look juvenile and antiquated.

Made in Japan went from signifying cheap to marvelous. Made for the Web is growing-up too. It is this evolution which guides my SEO highlights for 2012. I separate things to keep in mind by code, design and content.

Code – Keep It Simple

While Google likes to tell us they are very good at crawling and understanding imperfect code, I prefer to assume search engines are dumb and help them every way I can. Simple code is honest code. It’s also easy to parse and analyze. Just because you can AJAX-up a page with accordions and fly-outs does not mean you should. The more code on a page, the more things that can go wrong from spider access to browser compatibility.

Follow standards and get as close to validated markup as reasonably possible. Make it easy for search engines to spider your site. Validating HTML and CSS does not automagically raise your rankings, but it will prevent crawl errors.

At the same time, don’t insist on validation since some perfectly good code will never validate. Follow search engine recommendations to Make AJAXXML  and Other Code Crawl able.

Make your CSS class and ID names obvious, especially for section div tags. Again, Google tells us they have gotten good at identifying headers, sidebars and footers. Part of that is almost assuredly knowing the most common div names.

  • Make it easy on Google and Bing by naming your header div header.
  • Name the CSS ID of your right sidebar div right-sidebar.

Why would you name a CSS Class xbr_001 when you can name it navigation? At the very least, it will make life a lot easier on your SEO team. They have enough work without the need to translate ambiguous naming structures.

Reserve h# tags for outlining principal content. I am amazed at the number of big brand websites that still use h# tags for font design. Tell your designers that h1, h2, h3, h4, h5 and h6 are off-limits and reserved for content writers and editors.

The only exception to this should be if your content management system uses h1 tags to create a proper headline. Embargo h# tags out of your headers, navigation, sidebars and footers too. They don’t belong there.

Web Design – Less Navigation Is More

Look at the Zen like efficiency of any Apple product. Steve Jobs was ruthless about eliminating the unnecessary and achieving clean Bauhaus efficiency.

By contrast, too many websites, especially enterprise sites, try to be all things to all people. Their administrators or managers fear they might miss out on a conversion for lack of a link.

Websites should have clean vertical internal linking. Every page should not link to every page. You do not need a site-wide menu three levels deep. As long as people feel that they are progressing toward their goal or the useful information they seek, they will click on two, three or four links to get there.

Look at your website analytics. Which pages receive the fewest visits? Are any in your navigation? If no one uses a link, why does it to be there?

A website’s most widely visited pages tend to be close to the homepage. Review your categories and sub-categories. Can you eliminate whole categories by merging or reassigning content? For example, does the management team need its own category or can you move it into the About section?

This is not just about eliminating distraction. It is a way to increase the internal flow of authority (PageRank, link juice, etc.) to SEO hub pages.

Content – Engagement & Agility

Emphasize Community and Conversation. If your business depends on the Internet and you have the budget to hire one more person, consider employing a community evangelist. High rankings require authority. Authority comes from off-site links and, to an extent, brand mentions.

Earning enough links to make a dent in your SEO requires a continuous stream of link worthy content combined with forging and fostering relationships with people who create links or influence lots of others through online conversation. This requires a large commitment of time to work with writers and designers and to network. Even when decentralized, this rarely works without a strong empowered leader.

Get out of the sales funnel. The people you want to buy your products or services are not going to blog about your company or mention it on Twitter. More likely, they are peers.

A good exercise to undertake is ask each employee, if they could pick one professional conference to attend, what would it be? Then look for the session speakers on Twitter, LinkedIn and Facebook. Find which ones are active online and gauge their influence. Are people in your company qualified to write authoritatively about these topics or speak at conferences?

This is how to find content topics for the post-Panda Web, things people want to converse about and link to. For example, if you have a cutting-edge API team, an API development blog could be the key to higher domain authority.

Understand Social Technographics. It will help you to find influencers and create content that people will want to link to and talk about.

Embrace Agility

Realign your content generation and approval process so you can create near-daily web content and, if necessary, respond publically to something within an hour.

With Query Deserves Freshness, trending topics, news search  and simply because of how social media conversations come and go, agility is important for getting noticed and getting links.

Update Your Content

If your website has older articles that read like Wikipedia or a hardcover World Book Encyclopedia, swap out old content for new. In the future, Panda will not get leaner, it will get meaner. If you have reason to worry, start fixing it now. Do not wait and hope Panda will not see your low quality content. I want to be very clear here:

  • If you have decent quality content that provides real value, keep it whether it is SEO optimized or not. Yes, get to work optimizing older content doing things like selecting hub pages, optimizing text and cross-linking. But do not delete your old content.
  • If you have content that seems overtly advertorial, is cheesy or reads robotic because it is so stuffed with keywords, begin the process of writing one-for-one replacements and update your old content over time. For the old-time SEOs out there, this brings new meaning to a page a day.
  • If you have been hit by Panda already, I suggest removing your poor quality content, set-up 301 redirects to salvage the link authority, then begin rebuilding with high quality, link worthy content. Panda is a site-wide penalty. It is not going to go away until the offending content is removed or replaced.

Those are my 2012 SEO playbook highlights. In the past, content creation and link building were too separated. We had writers covering every long-tail key phrase possible while, in another room, link ninjas emailed and telephoned soliciting for individual links.

That model is becoming less and less sustainable. The Web is too big. Too many people contribute content. Social media offers an entirely new world of context. Today, SEO means finding an audience you can connect with, become a part of the community, give them insanely awesome content and reciprocate. This is the new SEO arms race.

Read source:


posted @ 2012-01-14 12:34 cnbarry 阅读(488) | 评论 (0)编辑 收藏

tool
http://www.linklicious.me

read
http://www.shareaschnitzel.com 
http://www.searchengineland.com

tojoin
http://forums.techguy.org
posted @ 2012-01-14 11:57 cnbarry 阅读(150) | 评论 (0)编辑 收藏

BY  

I’ll make it short and simple for you. Linklicious is not worth it. Stick with me for a minute or two and I’ll show you why. While pinging RSS feeds was for a long time a promising technique for indexing backlinks, today in the post panda world it’s just not enough. But that’s all that linklicious does. It turns your links into RSS feeds and pings them.

The panda update focused on getting rid of low quality pages. Whether they succeeded or not is a question for debate. If you look at the SERPs, we can see that for many searches the results are in many cases not as related as they were before the big algorithm change. If google identified your pages as being low quality, they may lose its rank or even get deindexed. It’s been said that if you have a number of such pages on your website, your whole domain might get a lower ranking score.

Back to the point, panda made it harder to index low quality pages, therefore your lower quality links, like profiles for example, don’t work as much as they did before. Let’s face it, if you need help indexing a webpage, it’s nature is most likely not of high quality. If it was a quality page it would get indexed by itself when spiders come crawling. So if we want our links to count, we need to index them. Linklicious should do that. But Linklicious just doesn’t do its job, or should we say, does it very “sloppy-ish”.

In order to prove this, I’ve ran a test for 45 days. I went rather crazy with a period that long, because links will usually get indexed no longer than 14 days after they’ve been crawled. If urls were not indexed by that time, chances are they won’t ever be, unless we give them some further push of course. For the test, I’ve picked one of the hardest links to index, profiles.

RESULTS AFTER 45 DAYS

There were 1903 live links that were built. And with the help of 45 days time google indexed 175 of those, using linklicious as an assistant. Keep in mind that most of those 1903 links were randomly interconnected, which helped with indexing even without linklicious.me. Some of those links were indexed by natural discovery of google spiders. So if we give a rough estimate, linklicious helped me index some 90 links in 45 days. I only have one word for this:

Blah.

Cheer your moods though, right now a fruitful test of a linklicious alternative is in process. In 2 days it managed to index almost as much as linklicious did in 45 days.

read source here:

posted @ 2012-01-14 11:42 cnbarry 阅读(276) | 评论 (0)编辑 收藏

今日google更新, 发现google site功能有变动。
以往site:url
返回结果为显示该URL相关的收录内容。
现在site:url
返回结果显示该相关内容以外,将url的地址格式自动去除协议、特殊符号等内容
既是转换为site: text text...
结论:
site功能依然保留,只是格式更新。
posted @ 2012-01-13 20:23 cnbarry 阅读(244) | 评论 (0)编辑 收藏

端口占用查询:
所有端口占用情况
netstat  -ano

正在监听的链接和端口
netstat -a

本地网络端口占用情况
netstat  -n

note:

NETSTAT Display networking statistics (TCP/IP)

-b           显示在创建每个连接或侦听端口时涉及的可执行程序。
              在某些情况下,已知可执行程序承载多个独立的
              组件,这些情况下,显示创建连接或侦听端口时涉
              及的组件序列。此情况下,可执行程序的名称
              位于底部[]中,它调用的组件位于顶部,直至达
              到 TCP/IP。注意,此选项可能很耗时,并且在您没有
              足够权限时可能失败。
-e           显示以太网统计。此选项可以与 -s 选项结合使用。
-f           显示外部地址的完全限定域名(FQDN)。
-n          以数字形式显示地址和端口号。
-o          显示拥有的与每个连接关联的进程 ID。
-p proto      显示 proto 指定的协议的连接;proto 可以是下列任
              何一个: TCP、UDP、TCPv6 或 UDPv6。如果与 -s 选
              项一起用来显示每个协议的统计,proto 可以是下列任
              何一个: IP、IPv6、ICMP、ICMPv6、TCP、TCPv6、UDP
              或 UDPv6。
-r            显示路由表。
-s            显示每个协议的统计。默认情况下,显示
              IP、IPv6、ICMP、ICMPv6、TCP、TCPv6、UDP 和 UDPv6
              的统计;-p 选项可用于指定默认的子网。
-t            显示当前连接卸载状态。
interval      重新显示选定的统计,各个显示间暂停的间隔秒数。
              按 CTRL+C 停止重新显示统计。如果省略,则 netstat
              将打印当前的配置信息一次。
posted @ 2011-12-26 20:51 cnbarry 阅读(246) | 评论 (0)编辑 收藏

It's me, mememe.    
posted @ 2011-12-26 12:46 cnbarry 阅读(97) | 评论 (0)编辑 收藏

仅列出标题
共5页: 上一页 1 2 3 4 5