Colo Colo de Futebol e Regatas

Colo Colo de Futebol e Regatas, usually known simply as Colo Colo, is a football club from Ilhéus, Bahia, Brazil. Colo-Colo was champion of the Campeonato Baiano, Bahia football league, in 2006.

Colo Colo
Full nameColo Colo de Futebol e Regatas
Nickname(s)Tigrão (Big Tiger)
FoundedApril 3, 1948 (1948-04-03)
GroundEstádio Mário Pessoa
Capacity10,000

History

On April 3, 1948, the club was founded by a group of sportsmen led by Airton Adami.[1]

In 1999, Colo Colo won its first title, the Campeonato Baiano Second Division, thus being promoted to the following year's first division. In the final, the club beat Fluminense de Feira.[2]

In 2001, Colo Colo competed in the Campeonato Brasileiro Série C for the first time. The club was eliminated in the first stage.[3]

In 2006, Colo Colo won its most important title, the Campeonato Baiano. The club defeated Vitória in both competition's stage finals, thus avoiding a final stage. In the first stage's final first leg Colo Colo and Vitória drew 1–1, and in the second leg, Colo Colo won 1–0. In the second stage's final first leg, at home, Colo Colo beat Vitória 4–3. In the second leg, in Salvador city, Colo Colo surprisingly beat its rival 4–2.[4] After the conquer of the championship, the coach and most of the team were signed to major football clubs in Bahia, like Vitória or Bahia.[5]

Achievements

Stadium

Colo Colo's home stadium is Estádio Mário Pessoa.[6] The stadium's maximum capacity is 10,000 people.[7]

The club's colors are blue, yellow and white.[6] Colo Colo is nicknamed Tigrão, meaning Big Tiger. The nickname is in the club's logo.[8] The club's mascot is a tiger, also depicted in the logo.[8] The club is named after the traditional Chilean club Colo-Colo, however the club's colors and home kit were chosen after the successful Argentine club Boca Juniors.[6] In 1948, the club's director, José Haroldo de Castro traveled to Argentina, and bought a Boca Juniors strip. The kit's design was then adopted by the club.[8]

gollark: Instead of recomputing the embeddings every time a new sentence comes in.
gollark: The embeddings for your example sentences are the same each time you run the model, so you can just store them somewhere and run the cosine similarity thing on all of them in bulk.
gollark: Well, it doesn't look like you ever actually move the `roberta-large-mnli` model to your GPU, but I think the Sentence Transformers one is slow because you're using it wrong.
gollark: For the sentence_transformers one, are you precomputing the embeddings for the example sentences *then* just cosine-similaritying them against the new sentence? Because if not that's probably a very large bottleneck.
gollark: sentence_transformers says you should be able to do several thousand sentences a second on a V100, which I'm pretty sure is worse than your GPU. Are you actually running it on the GPU?

References

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.