You are on page 1of 6

Fanatics Investment Banking Experience

Program Project 2

 Social Media Application:


Social media is a computer-based technology that facilitates the sharing
of ideas, thoughts, and information through the building of virtual
networks and communities. By design, social media is Internet-based
and gives users quick electronic communication of content. Content
includes personal information, documents, videos, and photos. Users
engage with social media via a computer, tablet, or smartphone via web-
based software or applications.
Among the five stages of the company life cycle, I believe the product of
this social media application lies in the First stage of commercialization
This is the second stage in which a company can look to raise money
from private equity investors. In this case, the company has already
launched its product in the market. In this case, the product is getting
traction but the company is yet to discover the perfect product-market fit
for its product. This money is broadly used to discover this product
market fit.
At this stage the organization should look to its customers to verify that
the innovation actually solves their problems and then should analyse
the costs and benefits of rolling out the innovation. The authors make
sure to note that “an invention is only considered an innovation [once] it
has been commercialized.” Therefore, the commercialization stage is an
important one,
The Caption recommending application gives Captions which are not
just for National Geographic! They convey vital information about who's
doing what, when, where and (sometimes) why. Solid captions paired
with interesting photographs can spark a reader's interest in a full text
story. Without captions, people draw their own conclusions about a
photo.

SWOT Analysis:

 Strength:

1. Give a Boost to Your SEO


One massive advantage of adding captions to your YouTube uploads is
that they let your videos rank higher in search results. Search engines
can’t understand and index videos (yet). They can, however, index the
captions embedded in them. This allows YouTube and Google to rank
your video for a keyword, even if it hasn’t been added as a tag at the
back-end. Studies have shown that videos with closed captions receive
a significant increase in views compared to those without.
2. Garner Wider Reach
Captions immediately make your video accessible to audiences, who
wouldn’t otherwise have been able to enjoy the content; for example –
people with hearing disabilities and those for whom the language of your
video isn’t native. YouTube especially, allows filtering for captioned and
subtitled videos for those who only want to browse through those
options.

3. Target Global Audiences


Adding subtitles to your video makes it available to people in foreign
markets, who although they don’t speak English, might be just as
interested in the subject-matter of your video. This improves the share-
worthiness of your content and lends it that extra boost when it comes to
virality. Not only that, search engines will index the subtitles and show
your video in search results in those languages. For niche content, this
can be particularly impactful, given the low competition for those
keywords in foreign markets.
4. Enhance Viewer Experience
Captions can significantly contribute to the viewing experience of the
average viewer, even if they’re not hard of hearing or non-native. Over
60% of video views happen via mobile, and captioned videos are very
useful for people watching from either very noisy environs like a train or
a mall or situations where silence is important – like libraries or office
spaces. Furthermore, captioning aids comprehension and memory
retention for viewers, especially when they’re watching highly technical
concepts being explained.
 Weakness:

1. Some foreigners still have to turn on the subtitles and captions for
helping them to fully understand the dialogues of figures in a movie
although they have arrived in an English-speaking country for such a
long time. But sometimes, they are distracting people since when
they concentrate on looking at the words on the screen, they seem
to forget the details shown on the screen. It will end up with watching
the news instead of enjoying the movies and performances.
2. Help to reach far more audiences. With the subtitles and captions, I
can enjoy the movie made by Japanese. With the help of these tools,
deaf or hard-of-hearing viewers can still find pleasure in the movie or
TV shows that they are watching.
3. Encourage the view. For those who don’t quite understand the
language used, without doubt, captions and subtitles can help them
to understand the details of the movies and deduct the meanings
between the lines in the written dialogues. However, the
disadvantage of it is that our precious time to do something else will
be reduced.
4. Opens the door to create shorter teaser videos that are intended to
spark the curiosity of a viewer so as to build excitement about what
comes next. without giving too much away. They are designed to be
a fifteen-second clip or a simple message such as “coming soon”
positioned above a date. Big brands often use these types of ads as
a frontrunner to their main campaign. However, during such a short
time, it is hard for the user to truly understand the content including
its disadvantages. Viewers may be misled and generate
misunderstandings about the products.
5. Capture the audiences and increase engagement at the beginning.
Similarly, if viewers clearly read the content, they thought that they
know what would happen and they may lose interest in continue
watching. Conversely, if they did not hear clearly the dialogue or
can’t understand what it means, they may insist on watching until
they know what is what.
6. Index the videos for SEO. Google search engines can obtain more
information from the text included in the captions and index the texts
to increase your keyword depth. beyond the title and description,
which may not be enough to ensure a good ranking. I can’t say what
disadvantages of this are? Maybe with your videos reaching more
people, more people would spend more time on the entertainment.
 Opportunity:

At the moment, the biggest usage of captions and subtitles in AR is


for assistive technology. Huge strides have been made to improve the
lives of those who are deaf or hard of hearing. 
One such example of this that has received lots of attention is the use
of smart glasses in theatres. This technology, dubbed Open Access
Smart Capture, was developed by National Theatre and will change
performances forever. Using AR alongside theatre projection
systems, theatres can fully integrate captions into the artistic
elements of performances. First, a 3D model of the stage set is used
to build a digital twin of the physical thing. Next, the digital twin is
aligned with its physical counterpart through the use of location
markers. Employing location markers ensures that the captions will
always be visible from any seat. Captions are then added to the
digital version and everything else is removed barring the captions.
Thus, when a user puts on the connected smart glasses, they will see
perfectly displayed captions. This means that deaf and hard of
hearing audience members will not have to constantly switch their
focus between the action onstage and the captions to the side. As a
result, they can enjoy a significantly more immersive viewing
experience. Providing access to culture and technology is a human
right and these glasses will contribute greatly to cultural diversity. As
more companies begin to invest in AR, the technology is developing
quickly.  Currently, AR is making massive strides when it comes to
live captioning. Universities have been working on using live
captioning with smart glasses to help deaf and hard of hearing users
navigate conversations. To start, a mic is calibrated to detect human
speech. The detected speech is then processed by a smartphone
sized micro-computer that the user can keep in their pocket. Finally,
the resulting text is transmitted to a display that superimposes the text
on the speaker. Although this setup achieves limited latency, the
captions are quite obtrusive. Furthermore, the user’s own speech is
also captioned which can create confusion. However, experts are
working tirelessly to perfect the technology.
AR technology is developing at a breakneck speed and the realm of
possibilities is continuously expanding as it does. Right now, AR-
based assistive technology is helping to improve the lives of many
and in the future this will only become more prevalent. Adding
captions and subtitles to technological innovations is a great way to
increase inclusivity. Fortunately, with Rev, it doesn’t need to be
complicated. You can get accurate and affordable captions on our
platform.
 Milestone 1:
Compositionality and Naturalness
The first challenge stems from the compositional nature of natural
language and visual scenes. While the training dataset contains co-
occurrences of some objects in their context, a captioning system
should be able to generalize by composing objects in other contexts.
Traditional captioning systems suffer from lack of compositionality
and naturalness as they often generate captions in a sequential
manner, i.e., next generated word depends on both the previous
word and the image feature. This can frequently lead to syntactically
correct, but semantically irrelevant language structures, as well as to
a lack of diversity in the generated captions. We propose to address
the compositionality issue with a context-aware Attention captioning
model, which allows the captioner to compose sentences based on
fragments of the observed visual scenes. Specifically, we used a
recurrent language model with a gated recurrent visual attention that
gives the choice at every generating step of attending to either visual
or textual cues from the last generation step.
To address the issue of lack of naturalness, we introduce another
innovation by using generative adversarial networks (GANs) [1] in
training the captioner, where a co-attention discriminator scores the
“naturalness” of a sentence and its fidelity to the image via a co-
attention model that matches fragments of the visual scenes and the
language generated and vice versa. The Co-attention discriminator
judges the quality of a caption by scoring the likelihood of generated
words given the image features and vice versa. Note that this
scoring is local (word and pixel level) and not at a global
representation level. This locality in the scoring is important in
capturing the compositional nature of language and visual scenes.
The discriminator role is not only to ensure that the language
generated is human-like, but it also enables the captioner to
compose by judging the image and sentence pairs on a local level.

 Milestone 2:
Generalization
The second challenge is the dataset bias impacting current
captioning systems. The trained models overfit to the common
objects that co-occur in a common context (e.g., bed and bedroom),
which leads to a problem where such systems struggle to generalize
to scenes where the same objects appear in unseen contexts (e.g.,
bed and forest). Although reducing the dataset bias is in itself a
challenging, open research problem, we propose a diagnostic tool to
quantify how biased a given captioning system is.

Specifically, we created a test diagnosis dataset of captioned images


with the common objects occurring in unusual scenes (Out of
Context – OOC dataset) in order to test the compositional and
generalization properties of a captioner. The evaluation on OOC is a
good indicator of the model’s generalization. Bad performance is a
sign that the captioner is over-fitted to the training context. We show
that GAN-based models with co-attention discriminator and context-
aware generator have better generalization to unseen contexts than
previous state of the art methods

 Milestone 3:
Evaluation and Turing Test
The third challenge is in the evaluation of the quality of generated
captions. Using automated metrics, though partially helpful, is still
unsatisfactory since they do not take the image into account. In
many cases, their scoring remains inadequate and sometimes even
misleading — especially when scoring diverse and descriptive
captions. Human evaluation remains a gold standard in scoring
captioning systems. We used a Turing test in which human
evaluators were asked if a given caption is real or machine-
generated. The human evaluators judged many of the model-
generated captions to be real, demonstrating that the proposed
captioner has a good performance and promising to be a valuable
new approach for automatic image captioning.

 For an augmented reality app, the customer acquisition cost will be


mainly driven by placing digital ads across a range of media to drive
installation. Boosting social media posts, google ads and having a well-
designed website can deliver a significant payoff. Although, the primary
cost the company may face is b-to-b costs like printing brochures for
students, sponsoring college events or providing promotional discounts.

You might also like