Hello,
welcome to this webcast, product accessibility evaluations. It's
being hosted by the
along with my cohost, Steve Jacobs. Steve is the president of
ideal at NCR, and its also the program accessibility manager for
NCR Corporation.
Steve thank you for helping us out today (beginning.)
JULIE CARROLL: Hello. And welcome to this webcast. Product accessibility
evaluation. It's being hosted by the information technology technical
assistance and
training center, or ITTATC as we call it. My name is Julie Carroll
and I'd be your mod raid to go along with my cohost Steve Jacobs.
Steve is
the president of ideal at NCR, and is also the program accessibility
manager for NCR Corporation. Steve, thank you for helping ut out
today.
STEVE JACOBS: Well, thank you for inviting me to cohost this panel
with you. I wanted to commend ITTATC for pulling together this
panel to discuss issues that are
very important with regard to developing accessible electronic
and information technology.
JULIE CARROLL: Thank you. The topic today, as I said, is product
accessibility evaluation evaluation. With the implementation
of section of the Rehabilitation Act, and section
of the Telecommunications Act, both of which of course have disability
accessibility requirements, a number has -- of issues have surfaced
surrounding
testing of products for accessibility. And today we will be talking
about some of those issues, including how to verify product accessibility,
how to make
procurement decision, and who should be responsible for product
testing. We have an excellent panelists, and they will be discussing
these topics as well as some of their own
approaches and recommendations in the area of product assessment.
But before I turn to our panel, let me give you an overview of
the format for today's Web tas. -- webcast. We
will spend approximately the first hour hearing from the panelists.
We will run from to o'clock eastern standard time much the second
hour we will open it up to the audience
and we look forward to many questions from you. There are four
ways you can submit questions to us. You can phone in your questions.
At --,
extension . You can fax your questions to us at --. You'll also
see on the screen, there
is an e-mail link and also a chat link and you can commit your
question -- submit your can hes that way. You can do it at any
time during the webcast.
I'll give the numbers out later. If you have technical difficulties
during the webcast, you can call and talk to one of our engineers
at --,
extension . And please tell us where you are from when you submit
your questions. I'll turn it over to our panelists. Our first
panelist today will be Dr. Dennis Folds. And
Dr. W. Bradley Fain. They are human factory specialists from the
Georgia Tech Research Institute at Georgia Tech, in Atlanta.
Dennis and Brad have developed a product assessment methodology,
and they did so in response to requests
from some of their clients. And as a result of the demand that
we heard from companies and also others who have to make procurement
decisions, ITTATC went
to Brad and Dennis and asked them if they would modify their product
so that it could be put into the public domain and used by anyone
who might be interested in this type of approach.
It is one of many types of tools of this nature that ITTATC hopes
to highlight for those of you out there who are interested in
testing
of products for accessibility. Brad? Welcome.
W. BRADLEY FAIN: Thank you very much. I wanted to start off today
by stating that the problem of accessibility evaluation is a complex
one and it is difficult. It is
a difficult topic. Unfortunately, the time allotted today, I'm
not going to be able to go into a lot of the important issues
in sufficient detail, however, materials will be
available on our website if you choose to go to them. And I encourage
everybody one to call in and ask the questions if they have them.
The methodology, like I
said, that I will present today is very complex, but in general
I don't think it is anymore complex than the accessibility field
in general. The main purpose
for this portion of the presentation is to present our methodology.
I'll talk about lessons learned, and open a forum for discussion.
What I'd like to talk about today
is the generalized methodology for accessibility evaluations that
we developed at GTRI. In response to an increased attention in
--
that accessibility has received in the past year, several of our
customers approached us about measurement of abscess built. And
they were interested in measurement of accessibility
in both prototypes that are in development as well as existing
product lines. So we developed specific test methods for these
customers to meet their needs. As
Julie alluded to, ITTATC approached us and asked us to generate
a generalized methodology to serve as a framework for the development
of specific evaluation
methods. It was designed to yield very detailed data to support
both product design and collection of data that would be
useful to a procurement officer when making compliance decisions.
The methodology is self is task based. We felt very strongly that
any assessment
of human performance requirements needed to be task based. The
methodology calls for a thorough checklist evaluation as well
as user in the loop testing. Used alone,
both checklist evaluation methodology and user in the loop testing
methodologies have substantial short comings. However, combining
the two methods into
a single methodology allows the consumer of the test data to have
a more complete picture of the product's overall accessibility.
Using the methodology, a method
of tasks may be tailor Toddrded -- tailored to a specific product
or test objective. The way we define test objective is both in
terms of the
user being looked at, the user population under consideration,
and the specific component of the device that is being tested.
For instance, an effective test objective
would be measure accessibility or evaluate the accessibility of
the front control panel of X, Y, Z, copier for low vision users,
or users with
low vision. The first part of the method that I'll describe today
involves the selection of applicable checklist items from a database
of items
that we collected. We collected over checklist items from sources
such as section , section the accessibility literature
and other human performance literature. We then classified each
item ak covereding to the types of guidance offered. For example,
some checklist items clearly only
applied to documentation. So we created a special category for
product documentation. We further divided the checklist items
according to specific user categories.
For each evaluation, the evaluator selects the subset of items
from the list of checklist items that are applicable to the objective
of the test -- of
the product that is being evaluated. The evaluator then determines
if guidance offered by each checklist item is met by the product
and renders a pass or
fail judgment. Now, the actual importance of the pass or fail
judgment depends on the evaluator's comments. Rendering a fail
judgment
doesn't really tell us anything that the designer can use, other
than that specific component did not pass a specific guidance.
If only a portion of the user interface
failed to pass the checklist item, then the evaluator notes the
reasons for the failure and notes what portion of the checklist
of the user interface that didn't pass.
Some of the items are purely quantitative and straightforward
to measure, and involves using a force meter to measure button
press of force. Involves using a measuring device
to measure keyboard height, for example. Other items require some
professional judgment to determine if the requirements are actually
met. It is of critical importance then that
the evaluator be very familiar with the requirements and to some
extent the intent of the checklist item. Now I'm going to ask
my colleague,
Dr. Dennis Folds, to explain the second portion of the evaluation
that involves performance of a task analysis and user in the loop
testing.
Dennis?
DENNIS FOLDS: Thank you, Brad. A key component of this task based
approach is first to conduct a task analysis of the use of the
device that is being tested. The first part of a
task analysis is to identify the task that users can perform and
one can use several sources to identify these tasks, if it's an
existing product that is being used by
a community of users somewhere, you can certainly observe users
and note the tasks that they perform. Technical documentation
is also another source of test. One thing to avoid
in using technical documentation as a source of task, is the tasks
identified should not simply be exercises the functions of the
device,
but rather the tasks should be stated in terms of what users have
to accomplish using the device, rather than just merely exercises
as functions. For products that do not yet exist or have not
yet been fielded, another way to identify tasks is to design scenarios
in which the various way the devices would be used by ultimate
users are
described in some detail. After identifying -- producing a list
of tasks in which the tasks of interest have been identified,
the next step is to prioritize those
tasks. Some people resist the notion of prioritizing the tasks
altogether, but I think most people agree that at least two levels
of prioritization is
appropriate. There are some tasks that I properly seen as common,
every day uses of the tasks, that any user would be expected to
be able to perform in order for the device
to be considered accessible. There are also tasks that might only
be performed once or rarely in the life cycle of a product, things
that are set ut when
the product comes out of the box, or major items that might be
performed and we might expect the user to
perform them. There might be something in between the extremes
that use advanced features of the device that are not commonly
used, but might be used for special circumstances. We recommend
allowing the three levels of prioritization to be the tasks to
be identified, each
task individually. After the tasks are prioritized, they are used
in developing the method of test, so the user in the loop testing
can be conducted so the users perform some of the
tasks. It's it's a case that the number of tasks on the task li
is larger than -- one would not include all of them in the user
in the loop testing. Doing the user in the loop
testing is also something that some people question. But, we feel
that it's very important when there are human performance requirements
that might be met that user in the loop
testing be conducted to illustrate and to verify compliance with
those human performance requirements. When you do user in the
loop test, there are a number of issues that have to be addressed,
not the least of which
is the selection of the users that will be used in the testing.
If I -- it's always an issue when the user in the loop testing,
you get a representative sample of
the population interests. This is true when dealing with accessibility
issues, particularly. It's easy to come up with categories of
users that you are interested in, but it's more difficult to determine
if you properly
sampled the range and variability of level of capability in that
user base. Nonetheless, it must be done. So one does one's best
to develop
a representative sample of users representing the population of
interest. In the user in the loop testing, the emphasis is on
task performance, rather than user comments,
although user comments are certainly welcomed and are useful.
The technique that we recommend in user in the loop testing is
to measure their
performance, to ask users to do what we call key event reporting
reporting. So rather than a continuous thinking out loud process,
they report when
they have trouble reporting a particular step or figuring out
what to do next. It's also the case that the user in the loop
testing often involves getting subjective ratings of whether a
certain
task can be performed acceptably, we recommend a four point scale,
going from completely unacceptable to completely acceptable or
satisfactory for
those devices. During user in the loop testing brings a number
of issues with it, not the least of which of course is the sampling
issue. But beyond that, it's also difficult,
in some tasks, to define objective performance measures, and in
those cases it may be possible to only rely on the user's subjective
evaluation evaluation. Other
issues with our methodology include the fact that it's best performed
by a person very knowledgeable of the standards and of the technologies
that are being evaluated, probably
more difficult for the untrained user people unaccustomed to doing
these evaluations to perform. It's also the case that it may be
time consuming. As it's time consuming, then the cost of using
this methodology maybe of concern to some users. Brad, do you
have a wrap up?
W. BRADLEY FAIN: Yes. In summary, the methodology that we described
can be used for both formative and summative evaluation. A formative
evaluation is usually performed early in the design process to
test prototypes and design concepts. The feedback generated from
a formative evaluation can provide a wealth of knowledge to designers.
It can help identify the source of accessibility
issues, and provide feedback of those issues to the designers,
and it can help designers understand the impact of potential interventions
that they might have. In
contrast, the summative evaluation is a formal evaluation of a
product, at or near the end of its design cycle. A summative evaluation
is designed to render pass or fail
judgments of design features. The results of a summative evaluation
can help designers or procurement officials for that matter determine
if accessibility requirements have been met. By
far, though, the most effective use of the methodology that we
have just described or any accessibility evaluation methodology
for that matter is the early integration of testing in the design
process process. By testing early, design flaws can be identified
and remedied prior to the product hitting the market. By combining
the utility of user in
the loop test, with a thorough checklist evaluation, I believe
we have an effective and flexible methodology for evaluating accessibility.
Thank you, Dennis and Brad. I have a couple of questions. First
of all, you're right. It sounds very complex. Is this something
that anyone can use and are
your materials documented in such a way that someone could learn
to do this themselves, or would they have to hire someone, experts
like
yourselves to do this for them?
W. BRADLEY FAIN: It's always the pocket that if they don't have
the in-house resource, that they need to go outside to get the
expertise. However, in conjunction with ITTATC we are developing
training materials
for this methodology that will be available. People will be able
to access the methodology on the Web. It's purely available. It's
not just something that we have any restrictions to distribution
of. However,
I would suggest that, in the -- if we were to do this evaluation
methodology effectively, to have background in dealing with
user testing and a thorough understanding of all the checklist
items being evaluated.
I have a question. For Dennis Dennis. You had mentioned
in your assessment methodology that you identified tasks, and
you don't
really address the functions of a device. But, rather, the task
completed, which I think is great. That's performance based, rather
than prescriptive methodology.
But, with a technology evolving the way it is, we are dealing
with really a global marketplace. If you take a web-based application
in
E-commerce, for example, what would you recommend be done to get
representative samples of your population together, when it could
be anyone, anywhere in the world? It's
easier to do that if you're working with a device that you can
feel and touch. But, the type of interface that is accessible
to millions of people all over
the world can change dynamically. You probably see that today,
where you go to your ATM, and what you see today is different
than what you might see tomorrow tomorrow.
The question --.
I'm sorry. The question is: How do you factor into your methodology
the variables I just described?
In the user sampling question in particular, it's important to
identify the attributes of users it's important to evaluate in
this sort of methodology. So, for example, you
mentioned the question of worldwide usage, and of course with
worldwide usage the lack of language or perhaps even literacy,
English or even
literacy, might be an issue. So, if you are developing an interface
that is not dependent on English or even on literacy, as its basis
for task performance, then you
need to create the conditions of performance in your user in the
loop test, so that language is not available to support the task
performance, and then see if the task can be per
forpd. -- performed.
It's also the case that even in this country, one in large cities
does have access to a very broad range of people that don't have
English as a first language or perhaps also have some
literacy issues.
STEVE JACOBS: Thank you.
JULIE CARROLL: I have a question for you, Brad. You mentioned
pass fail kind of grade. Just curious, you know, as a blind person,
I might be able to complete a task, but
what if it takes me minutes, where it takes other people a few
seconds, would you consider that a pass or fail? How does that
factor into your analysis?
W. BRADLEY FAIN: Well, the pass fail judgments themselves are
not rendered with user in the loop testing. That is sort of a
separate issue. What you have for the pass fail judgments are
checklist evaluations, where
a professional evaluator or somebody that is responsible for the
test data will render that judgment. And that's based on objective
measures or subjective measures if it
requires professional judgment. However, the question that you
asked has to do with a judgment as to whether task performance
can be completed or not. We have three
categories of data in that case. We either say that the task can
be completed without difficulty, meaning the user was able
to perform the task with no apparent obstruction or difficulty
in performing that task. The user was not able to perform that
task, meaning they couldn't complete all the task requirements.
Or, a middle ground saying that the user had difficulty in performing
that task. And that information is rolled up into the results,
to show the designer where the -- maybe
the speed bumps in the interface actually lie. So we break that
ouchlt we don't render a pass/fail judgment in user in the loop
testing, however.
STEVE JACOBS: We have a question from the audience. Thank you
for calling in. This is from Austin, Texas. And I'm going to address
this to Dennis and Brad. If a procurement official
determines that product A meets six out of the ten applicable
requirements, and product B meets another set of six out of the
ten requirements. How does the official make a determination as
to which product is more accessible? That's not too hard of a
question. I'll leave it
up to either one of you to address it.
Go ahead. Using this methodology, it's possible to compare levels
of accessibility of different products. But it's also possible
that one would be unable to distinguish whether one product was
more accessible
than another. In the case that you just cited, six out of ten
versus six out of ten. That sounds like a tie in the way the question
is worded. However, using the task based
approach, one can identify, for a specific user population, the
percentage of the priority one pass that -- tasks that can be
accomplished for one product versus another. Priority tasks are
those tasks that we expect any user to be able to do at any time
using this product. So I think that
would provide a basis for a procurement official to distinguish
between two products even if they were apparently tied in terms
of number of different attributes.
STEVE JACOBS: Thank you.
JULIE CARROLL: I have one more question before we move on. Thinking
from the perspective of let's say a procurement official how would
they use a report from this type of methodology?
Are they going to get percentages or grades or are they going
to get a report that is going to help them make a procurement
zigs?
W. BRADLEY FAIN: Absolutely. If the test is designed to measure
compliance with section , then the checklist items will be very
much oriented in that fashion according to the standards.
So we will render judgments as to whether we believe that it comes
close to passing or failing or whatever. And we will note specifically
what component of the product
failed to be, in our evaluation, failed to meet the requirements
of the standards. In addition to that, they will have task based
information
in order to help them, perhaps, Judge the effectiveness of one
product versus another. So they will have both types of information.
So I think, you know, in summary, I think
that the combination of that information will help procurement
officials make better decisions. Just because the product A, again,
meets six out of the ten,
and maybe product B meets six out of the ten as well, deciding
which one is a better product will be much easier using this methodology.
JULIE CARROLL: Thank you. Our next panelist is Terri Youngblood,
and tirry is the vice president for technology of The Accessibility
Group. Or TAG as it's commonly
referred to. And she is also president of the accessibility systems,
Incorporated. And Terri is a nationally reknowned expert in the
area
of accessible information technology. Terri, tell us about TAG
and what your approach will be.
TERRI YOUNGBLOOD: Thank you, Julie. As you said, my name us Terri
Youngblood. I'm vice president of technology for TAG and it's
an acronym. The ak test built group is what
it stands for. It's a nonprofit organization. And we are dedicated
to the proposition that technology should be accessible and usable
by all people with a full range
of abilities and disabilities. The board of directors of TAG is
very important to point out, because we are made up of talented
veterans and experts in the accessibility
disability community. Our executive director is David Bomick,
formerly with Microsoft accessibility, disability group and
is a key thinker in the arena. Larry Skadden, year veteran in
the accessibility field. He was at the national
science foundation, programs for persons with disabilities. And
myself, as she mentioned, I'm the president of accessibleible
systems incorporated, which is an information technology accessibility
consulting firm. Working for the last years in cooperation with
government and industry, helping to make products accessible and
usable by the greatest number of people people. And also,
Modi
Cradamer, who is the -- our director of communication, who is
also the director of the award winning documentary enable for
people
with disabilities, which you may have seen. It was produced by
Microsoft to help understand usability issues of people with disabilities.
And he has also been a long time advocate to fortune
companies on accessibility issues. Peter blank, director of law,
health and policy, at the University of Iowa. So together these
five thinkers came
up with the group TAG, The Accessibility Group. We are made up
of independent third-party testing facilities, located nationally,
to provide
a valuable service. Service to industry, service to government,
as well as special interest groups, helping to achieve IT accessibility
conformance, accessibility with
the provisions of the technical standards of . We want to help
meet the growing need for industry. We want to help industry with
, with the technical standards out, we find that industry -- IT
manufacturers don't necessarily have the expertise, in-house,
nor do they employ the is
individuals with disabilities that have the testing techniques
or the testing skills in order to test the product to determine
whether it's accessible sibl. Julie, I've been in this
position for a while where industry comes to me and says Terri,
we really want to do the right thing. We want to make our product
accessible but we don't know how. That's where TAG steps
in and helps. A product can be submitted for evaluation. The product
can be pre-released or already released product. And TAG has staffed
our
national labs with qualified individuals with various types of
disabilities and skills as well. Also, engineers are present to
-- as a backup for the
standard testing, will serve as a quality assurance checkpoint.
We are also there to help government. Procurement officials and
requiring oh figures are left with the responsibility to gather
market research in order to make the most accessible purchase.
Market research can consist of
marketing materials or it can consist of going to the -- to the,
by accessible database and getting a voluntary product accessible
template,
it's referred to as the VPAT. And they may search for other techniques
or other market research. The problem we are seeing now is that
these requiring officials are looking to
the VPAT and they are finding that the VPAT is somewhat inconsistent.
Manufacturers, either the manufacturer fills out the VPAT and
they don't really have the skill
and understanding to understand the standard and, as you said,
what it's trying to achieve. And so a lot of times the VPATs
are filled out and almost like a marketing format. And interestingly
enough, looking -- and the concept of the VPAT is an excellent
idea and
we support it fully. But when you look at it the way it is now,
because there is a lack of knowledge out there in the industry
about accessibility, you find you can take, if you visit the by
accessible data bairx you can have one product and we have done
this, actually, one product two different manufacturer, and you
look at the VPAT, and they are filled out apparently by people
that may or may
not know exactly what was intended. So you have one standard that
one manufacturer says we don't apply to this standard, even though
it's the same product. The next company
says we apply to it and here is how we apply to it. So it's very
inconsistent. So as a procurement official or requirement official,
would go -- either the marketing material or
the information from the VPAT VPAT. They pull two templates
down and they look at them next to each other, and they say I
can't tell which one is more accessible.
So TAG is here to provide standardized reports that are -- have
the same look and feel and that procurement officials can easily
compare and
contrast from one to another. We have several different types
of services TAG offers. First one we call the audit, which is
actually testing and
providing independent review based on the technical standards
of section . So the results will be in a similar checklist format,
with an explanation of each of
the standards and how that product meets them. We also have --
we go above and beyond, because of our mission and our mission
is to educate and employ the
greatest number of individuals with disabilities disabilities.
By removing the barriers to access the technology. We also do
what is called an accessibility audit. That can be ordered by
industry, if
they want to look at a particular feature of their product, or
say that they met the standards and they want to go above and
beyond, they want to take that extra step to make the product
usable by even more individuals. Unfortunately,
as we know, even if you follow the technical standards, there
hay be some way, some barrier that still exist, let's say. So
the accessibility
audit can go above and beyond. And in that we insure testing,
compatibility testing and user testing. So compatibility testing
with assistive technology, assistive
technology that is used by individuals with disabilities. So often
manufacturers in today's situation may or may not be testing their
products when they complete the VPAT,
the existing template. So we think it's very, very important not
only to determine will you met the standards, but to also go above
and beyond and to test with assistive
technology and do end-user testing as well. So we can serve as
a service to those folks in the industry who don't have the knowledge
and expertise expertise. And they can take the report back
and they can
use it as -- in their R and D development, in putting accessibility
into their design process. We believe that industry should --
industry manufacturers of technology should also be testing internally.
TAG doesn't want to replace that by any means. But we should be
the
spot check, so to speak. Without simplifying it to the audience
members who may not be as technical, if you write your own paragraph
and you want -- and you read over
it and you read over it and you think everything is just fine,
but you also maybe will give it to your editor and I'm sure he
will find something that you didn't think about. Or let's take
it a step further.
Give it to the person who you were intending to be your audience
before you release it. And say give me your feedback. Did I get
my point across? So that is the concept that we are look at.
And the role. So we are looking at industry's already good efforts
in this arena. And we have what we call rapid review. And that
is also for industry or government. And that
is kind of a scaled down view of that. And one use of this review
would be government procurement officials or the CIOs of different
government agencies looking for a particular product product.
For example, maybe they are looking to buy a new e-mail system.
And they go to -- they go to the by accessible
database maybe the things they are looking for are not listed
in the VPAT. So then they could hire the TAG labs to do a rapid
review of each of the software packages
in consideration, and we would do a audit against that. We call
it rapid review, because it's a snapshot at the top feature, to
give them a general idea of where to go
from there and then where to seek more market research on that
individual tool. We have -- it's very important that TAG
-- TAG feels very important about industry and government, the
disability community, working together to insure that
is a successful transition. Thank you.
JULIE CARROLL: Thank you, Terri.
STEVE JACOBS: Thank you.
JULIE CARROLL: A couple quick questions. Are you suggesting that
this type of testing should be mandatory?
TERRI YOUNGBLOOD: No. Absolutely not. It's voluntary. Each organization
may have expertise internally that they feel can provide testing
adequately. But others don't. And
so no, it's completely voluntary.
I had a question for Terri. By the way, for those of you wanting
the call in numbers, should I give the call in numbers, Julie?
JULIE CARROLL: Sure.
STEVE JACOBS: If you have a question or would like to fax your
question in to us, the phone number is: --. Extension
. That is for a telephone call. For a fox, it's --. And on
the Web page that you're using to access this program, there is
also an e-mail link and a chat link. And we have another -- did
you have a question, Julie? We
have one from Los Angeles. And hello, Los Angeles. What is the
best, and I'm addressing this to Terri, what is the
best procedure for the validation of a particular methodology?
TERRI YOUNGBLOOD: What is the best procedure -- read it again?
Sorry.
STEVE JACOBS: I wish I could tell you what the writer intended
to say. What is the best procedure for the validation of a particular
methodology.
JULIE CARROLL: Good question.
STEVE JACOBS: Probably a testing methodology. If a company has
an opportunity to test their product this way or this way or this
way, how do you value a method of testing?
TERRI YOUNGBLOOD: Well, I think the end result -- the two results,
one being is the application or is the product conforming to the
technical standards? That
being your goal. The larger goal, though, is the testing methodology,
the goal to achieve accessibility, usability by
the target audience. So I would -- the best methodology is to
insure that, when you're testing the product, is your end result,
conformance with the standard and
accessibility of the product.
STEVE JACOBS: Is there a way to document that ahead of time, to
let somebody look at before they select a particular methodology
to evaluate their
products?
TERRI YOUNGBLOOD: Sure. With the end result being, for example,
our different views, section audit would have a testing methodology,
performance
technical standard by technical standard, feature by feature.
They would have one methodology. Now, if someone wanted to go
above and beyond the technical standards and wanted to find
out is my application accessible via a screen reader, it would
be a different set of technical. So it can be documented and they
can choose between the different options.
STEVE JACOBS: I have to make a correction in a phone number. Apparently
somebody in the studios is getting a ton of telephone calls and
they were not supposed to. I gave out the wrong extension. It
is not what I gave you. And I won't
repeat it. But the correct extension is extension . The whole
number is: --
and there should be a template to evaluate how accurately I'm
doing this. Sorry. I had one question for Terri. This is
my question. If any company had a choice of hiring a third-party
entity to, and in their expertise to test a product,
why wouldn't they just simply want to hire someone with the same
expertise as the people in this third-party entity, and bring
that function in-house, so that they
could test a myriad of different systems using that expertise?
I know your chairperson, Dave Bolnick came from industry and I
have a high
degree of respect for him. Hi Dave, if you are listening. Terri?
TERRI YOUNGBLOOD: We encourage it. It's not something that we
discourage. Our group of contacts, so to speak, that are
doing testing for TAG labs are a variety of the individuals you're
talking about. Our relationship with testing, testing groups around
the country are
just the expertise that you are referring to. If you want to bring
-- if an organization wants to bring the expertise in-house, we
think it's a great idea. It helps them to
instill the accessibility into the design process.
STEVE JACOBS: Thank you.
TERRI YOUNGBLOOD: You're welcome. I wanted to add something, if
you don't mind, not necessarily to that question. TAG's report
leaves two choices for industry. First of all, it is not a certification.
That is a question that people often ask me. Does TAG certify
your product? And the answer is no. We have what we call a logo
program. If a manufacturer
gets a report on a product, they can keep the report private.
And that's important to know, because initially industry was a
little nervous. We don't want all
of our information to be given to the public. What if we need
to make some changes to it. And we said that's very important.
So the report can be kept private and industry can take those
findings back to their developers
and product managers and say look, these are some of our weak
points. How are we going to innovate? And that's what industry
does best. So this report will help
industry innovate on how to make their products more accessible.
On the other hand, it can also be made public. So, if a product
is very accessible, and
it meets the standards, a manufacturer can have the TAG logo on
their product and the logo does not mean certified or not certified,
compliant or not compliant. And it clearly states that this product
has been tested by an expert independent group. And what that
will indicate to a procurement official or to a requirement official
is
that, Hmmm, there is more market research here for me available
other than the marketing materials of the manufacturer. It kind
of puts the manufacturer in a
sticky situation by filling out the VPAT. It says Hmmm, I don't
want to tell everybody about a couple weaknesses I may have. So
they have an inherent conflict of
interest by filling that out. So this logo program will give them
an opportunity of demonstrating that they have seeked an expert
evaluation.
STEVE JACOBS: Thank you.
JULIE CARROLL: Terri, are you not concerned that a logo sort of
connotes a seal of approval and could be misconstrued to imply
that something, TAG
has determined something to be accessible, when in fact, the report,
if you could get your hands on it, may show it's not accessible
at all?
TERRI YOUNGBLOOD: We are not concerned bit. Because the logo clearly
says it has been tested, and that means that information is available.
Much like a consumer's reports kind of that you. You know
if something was tested by consumer reports, the data is available.
That's what we are saying. I'm not concerned concerned.
But that's a good question.
JULIE CARROLL: One more question about testing. I assume that
when you talk about testing you're talking about does the assistive
technology that a person might need to use with it
work, when you use a particular product or can a person with a
disability use that product? I think there is a flip side to that.
And
sometimes when you plug in assistive technology to a product,
the assistive technology might work, but the original product
no longer works properly. Do you test both sides of that?
TERRI YOUNGBLOOD: Yes. We do. That's very important. Because it
is a compatibility between the assistive technology and the mainstream
technology. It's a very important
point you brought up. Because I commend industry, since the reauthorization
of industry manufacturers, because they really have been taking
an active role in communicating with the
assistive technology vendors, and they are much smaller, compared
to the big boys. So keeping up and keeping in communication is
very difficult. Industry has been doing a very good job of keeping
in touch. But TAG, by doing
compatibility testing, as well in our accessibility testing will
serve to boost that communication process even more. We hope to
keep an open conversation between industry and the
assistive technology vendors, hoping to facilitate even more communication.
JULIE CARROLL: Thanks. Junk junk thank you very much. Carol care
we are now going to move onto the next panelist. Kenneth Salaets
is with us today. He is the director of government
relations for the information technology industry council, here
in Washington. Ken has also been involved at a national level
in a number of initiatives relating to the implementation of
section , including the development of the Voluntary Product Accessibility
Template, or VPAT as Terri has referred to it. Ken, thank you
for joining
us.
KENNETH SALAETS: Thank you. I'd like to thank you and at particular
for the opportunity to participate in this webcast, the first
that I've had the pleasure to be in. And also thank Steve
Jacobs of NCR, one of the valiant members of our working group
at IT I, for his participation. The information
technology industry council. We have other companies such as NCR
IBM, Compaq, OAOL,
time warner. I could go on. It's quite a group of innovative and
progressive companies. Frankly, it has made my job far simpler
to have the resources available that we
do at IT I, as a consequence of these companies committing time
and effort and in some cases dollars, to
assist the program. We have been around since . NCR, the national
cash register, was one of the
charter members. We have gone through a lot of name changes, but
all in all we represent what I believe are the most opportune
issues, in one of the most opportune
industries as impacting the economy but making distinct differences
in the lives of all of us here and probably everybody that is
accessing this webcast. We
have -- cutting edge is a cliche, but if you wanted to talk about
cutting edge technology and cutting edge companies, I think these
are them. And it really affords an opportunity in Washington,
D.C. and around the world to have doors opened for you when you
are able to mention that you are working on these particular issues,
and trying to open the digital
network to all user, all individuals, worldwide. It's a challenge
that we have gladly taken upon ourselves, and we enjoy the opportunity
again to engage
in discussions of this nature. That brings us to section , how
did we get here? Red Dawson, president of IT I, made a personal
commitment a few years ago
to make sure that we were directly engaged on this issue. Our
companies made us very aware that this was of distinct importance
and interest to them as manufacturers and suppliers of technology.
And
he gave myself and my predecessor, John Gottfried wide berth in
trying to work with the companies and
government and advocacy groups to try to tackle the challenges
of enabling the government to identify which technology in the
marketplace would meet the needs of their --
Allen doesn't like the word end-users, but their employees. And
we took that on, working closely with Terri Weaver at GSA, for
example, in trying to
identify a mechanism that we thought would be convenient and useful
to procurement officials to take a snapshot of the marketplace
and try to ascertain what was available as of the
various effective dates of the standards to be able to begin to
address the needs of the employees, as well as private citizens
accessing government information. GSA started
working on something akin to a template at about the same time
we did. So we joined forces with Terri and her folks to try to
hammer out some kind of an agreement or
a document, which we then called the template, that would assist
the procurement officials in taking that snapshot and complying
with the federal acquisition regulation
that says they have to go out and do market research just as we
do in the private sector, to ascertain what is in the marketplace.
It wouldn't make any sense at all for a federal agency to put
together a
solicitation for products if, in putting that solicitation on
the street, as we say, nobody could respond with a product that
could meet the requirements. So, in recognizing, as the access
board did, as well as Craig Lugard, the chairman of the federal
steering committee on , that there is no on/off switch
that you can flick and suddenly everything is accessible. We started
working on what we thought was a transitional process toward helping
the agencies identify what was in the marketplace, but also helping
companies, helping businesses to provide information to find some
mechanism, and a useful, free mechanism, snd plying information
to the procurement officials. What we have
found so far that is the template, the VPAT, has proven to be
a fairly popular tool, notwithstanding some of the
comments you may have heard here today. We have, for example it's
available on IT I's Web Page. That is our acronym. It's available
at www.IT IC.org.
If you go to that Web Page and it's approximately in the middle
of the home page, you find an article that identifies accessibility.
If you click on that, that takes you to documents that explain
our position and ultimately to the template itself. I believe
Terri Weaver made it
accessible by the by accessible.gof site. And by the number of
hits that we are receiving, you find links on a number of other
pages throughout the government. I'm proud to say that the section
related documents on the IT I Web Page are still
receiving probably the highest volume of hits, and that is both
from the federal governments, state governments, from private
companies, advocacy and also foreign governments and
foreign industry. So we believe that it has been a useful tool.
We are happy to provide the service. Again, free of charge. And
we put it out there on the marketplace of ideas for use by both
government and industry.
Is it perfect? No. It's not perfect. Is anything perfect? No
No. There is really no panacea, no silver bullet on accessibility
and particularly on
determining whether a product is or is not accessible. Primarily
because you have to determine the individual needs of the individual
user. You may have fairly convenient labels
that we sometimes fall into, in using to describe certain types
of disabilities. But in reality, there are so many variations
and colors, so to speak, that
one solution which may be fine for people, may not serve the needs
of the th person. In fact, may not be effective in serving the
needs of the th and th
persons persons. So as a consequence, we encourage the access
board, through the advisory committee that they had beforehand,
again, Mr. Gottfried is serving on that board on behalf of
IT I, to build flexibility into the standards. The reasons being,
if you try to drive a stake into the ground and try to be very
explicit and descriptive
in describing what the technology has to do and how you have to
describe that product, companies will billed to that single point
and probably won't go beyond it, and innovation dies. One of the
advantages
of our marketplace and industry is if you describe what you are
trying to accomplish in terms of performance requirements, then
we will stumble and fall all over each other, throw our
engineers at it, R and D and everything else, to try to out compete
each other to come up with what we believe is the best solution.
As a consequence, you'll end up with four or five optionness that
marketplace and that's the very thing the government tends to
look for. The government prefers to buy or have multiple options
-- multiple providers for the simple fact that they can force
us
to compete for that business. We find the same thing occurs in
the realm of the people with disabilities and advocacy groups
in terms of looking for solutions. They also recognize that one
size
clearly does not fit all. And if we had a one size fits all world
from which to buy, there would be very little choice. And as a
consequence, very little innovation.
The template is a voluntary mechanism for providing information.
We don't prescribe how to use it. We don't -- in fact, there was
a prescription handed
back to usaq that rather try to take the standards themselves
and interpret them in a certain way, that rather we print verbatim
the
standard itself into the template. The Department of Justice had
a hand in that, in advising Terri Weaver and the like and that
information was pass odd to us and we accommodated that. Nevertheless,
if you paid attention to the access forum, the event that occurred
in Denver, there was quite a bit of comments and criticism about
the template and about the
experience that some government officials are having, including
contracting officers who of course were not expected to have to
use that, but unfortunately, in some cases, that
responsibility is being flowed down. And I don't envy their job
when they are trying to manage contracts on a daily basis. But
the intent is to -- this -- the
template is a work in progress. As you look on the ti IT Iwebsite.
That is the way we define it. We have fully intended to continue
to work with government, work with advocacy groups
and also the AT industry to find ways to improve this tool. In
fact, we entered into discussions with GSA, to try to generate
some dialog with the contracting officials
that have been using this tool to hear specifically from them
what concerns they have, what weaknesses they may perceive or
what additional tools they think might be helpful.
What we could do is not necessarily change the template itself,
but we would add additional resources on the IT! ITI website
that provides guidance
on how best to use it, with the hope that when business looks
at the template and decides that they will fill it out, thr that
there would be a degree of consistency in experience. So when
contracting
officials go out to the Web Pages of various company, whether
through the by accessible.Gov site or other Internet sites, that
they would have the ability to make more
comparisons in terms of what is being represented in terms of
the material. But ultimately, the companies themselves are responsible
for the information they provide, it's designed to be a tool.
It's not
an end all to be all. It's the beginning of a dialog. Very seldom
do drting officials particularly in large procurements go
out and make a decision based on a single piece of paper. In fact,
they directly engage with not only the sales force but also the
designers and engineers of the companies to
try to derive as much information about the product and technology
that they are contemplating. That is called competition. And the
government is effective at driving industry to a very
fair price for very effective technology and service service.
And if that were not the case, frankly, they would not buy it.
We believe in the farkt place, that the marketplace will
make distinct -- determinations on what is and is not an effective
product or an effective offer. You have seen the marketplace on
this panel right here. There are a couple competing testing
methodologies and groups here that are advocating their services,
and we think that is great. We think there ought to be as many
possible options in the marketplace as possible. That is the nature
of the marketplace.
Sometimes you develop what we refer to as cottage industries,
where a whole new industry or groups will spring up to address
a certain issue. That is the nature of the market and we applaud
that. Companies should be free
to test internally and to make presentations, representations
to the government, regarding the accessible features of their
products. We do that in save techlt we do that in electromagnetic
fields, we do that in the energy star program and the like and
we have yet to have anybody suggest to us, I believe, on a blanket
basis, that somehow we are being deceptive in terms of
those representations. Does that mean that everybody is as complete
as they can be? No. Does that mean some companies are better at
it than others? Yes. Do
some companies have more resources to meet it than other smaller
companies? I think that's a given. But nevertheless, crystalized
this marketplace.
It woke up the industry I think it's fair to say. It has helped
to generate an ongoing dialog between the IT and AT industries
that is growing more healthy by
the day and will even be engaged in further, next week, at CSUN.
And we opened the doors and the channels and increased the opportunities
to talk to people like Allen
and his group and other individuals that represent people with
disabilities. We are encouraged by this process. We applaud Georgia
Tech and its effort and
innovation that they have shown in trying to tackle this through
the education department contract. We have seen other groups that
-- such as accessibility forum also show sincere
effort to try to address these very important issues. Again, I
represent some of greatest companies in the world. They are highly
committed committed. I can run through a whole list of names.
I won't. Because I'm sure I
will forget somebody and I'll hear about it later. But I really
appreciate the opportunity that I've had to be involved in this,
because, frankly, I didn't pay a lot of attention to these issue,
until I
got this accessible device and now it sits on my nose, called
glasses. Because of my Vanity, the bifocals, you can't see. I
have come to appreciate the importance of this technology. Somebody
said to me that with the aging of the USpopulation, the world
population, that technology
is going to become less and less accessible to people. With all
due respect, I dismiss that. I believe that technology is going
to open doors, windows, ears, eyes, it's going
to create a phenomenal new world where we will be able to access
information beyond our dreams, even for those of us who assume
that we don't have a disability or at least
have not yet discovered it. Again, Julie, I thank you for the
opportunity. I welcome the opportunity to respond to questions.
We believe in partnership, whether it's with government, with
other
industry groups, with the AT industry, the advocacy groups
groups. And we encourage them to contact us through the website
or directly at --.
I'd be happy to respond as I can. But this is not the only issue
that I work, so I have to juggle Aother issues important to our
members. But we are committed to pursue this and
carry this forward as aggressively as possible and we welcome
everybody to join that effort.
JULIE CARROLL: Thank you. I have to say that working with ITTATC
has been great to seeing the enthusiasm that has crept into industry
for
accessibility. Thank you for your comments. I have a couple questions.
The voluntary product template, are you suggesting that it is
sufficient by itself that companies shouldn't
be expected to do anymore? What do you think about this third-party
testing?
KENNETH SALAETS: Third-party testing, we think, is a useful tool
to complement industry's own test, to complement government's
own research, to
complement the best practices that tends to develop in the marketplace.
One thing that I've come to appreciate in terms of government
contracting, and dealing with the government, is that contracting
officers are highly professional individuals who have a far better
understanding of technology than I think we give them credit for.
And the reason I say that is because I sat down with many of them
over the
years, the last years that I've been with ITI, and have found
that -- often I've gone into the meetings with misconception about
their understanding of not only the acquisition code but also
of the nature of our industry and the business itself. I think
that some of the risks of third-party testing, if it's viewed
as sort
of the panacea or if it's shopped as the sole solution to enabling
procurement agencies to you know, meet their
requirements, their obligation to comply with , is that in and
of itself I think they tend to have a -- there is
a tendency for third-party testing to interpret the standards.
And while we all try to tackle the standards, try to read in between
words, in
between lines, to try to ascertain exactly what the access board
was say, and the more importantly trying to meet the needs of
the individual employees themselves, I think it's important
that we not find ourselves restricted in terms of the ability
to respond with innovative products to compete for that space.
And there is some risk that third-party testing can start to
become interpretive and there by limiting in terms of the access
standards.
STEVE JACOBS: I have a question for Ken. Before I do that, though,
before this broadcast, we had a bit of lunch and I was talking
to Ken. I hope I don't embarrass you. But Ken's daughter is watching
this webcast. Her name is Karen. And Ken didn't ask plea to do
this, but I thought I'd let Ken say hi to Karen. Go ahead.
KENNETH SALAETS: Her name is Kera.
STEVE JACOBS: I do have a question for you, and it may be one
that I think a lot of people in the audience would want to ask,
but we didn't receive this as a question from the outside. You
represent a very diverse group of companies who
compete with each other. Do you find it difficult to find ways
to get all of the companies to agree on any one particular
way to do things or be at the VPAT or third-party testing, or
anything anything? How does that all work work?
KENNETH SALAETS: It's very interesting -- it's a very interesting
world in which to work, for a number of reasons. One is you often
-- not in this issue, not
in , but let's talk about some other issue, for example, you may
find that if you had companies at the table, you might have in
general agreement and two violently
opposed. And the challenge in the nonprofit industry, association
environment, is to identify common denominators among all of the
companies, and then try to build
from there. What we tend to do is if there is an issue, what I
tend to do, I'll speak for myself. If there is an issue where
I don't believe we can establish consensus, then we might simply
have to agree to disagree,
and not work that issue. In section -- with section and accessibility,
it has been the complete opposite case. And one reason is the
commitment of all of the individuals that
our member companies have put at the table. I would put in a plug
for Microsoft, Laura Ruby. She is an outstanding chair,
hello, Laura. Tell my boss. And she is an excellent facilitator.
And we have also other exceptional individuals who bring their
ideas to the table. They will fight for their ideas. They will
argue every possible way they can think of. And we tend to be
a very collegial
group in the sense that we all are trying to pursue I think the
same objective. In some case, when I hear the individuals I work
with I'm trying to ascertain, are you representing your company,
are you
representing people with disabilities? Because they are so closely
knit within the various communities that they serve, and of course
committed to their companies as well, and their ideas, that I
think it is really facilitated
the opportunity that we have had to work with the different groups
and to develop the rapport that I think we have. As Julie said,
that we finally discovered, I would like to think that ITI may
have been ahead of the game in terms of having that dialog and
particularly the individuals, many of whom predated and predated
, for that matter, working on accessibility. A number of
our companies were engaged in this business for , years. It's
not necessarily something that just fell in their laps. But then
again, section , or the amendments, I should say, have again crystalized
that marketplace and they focused all of our attention on this
important issue.
STEVE JACOBS: Thank you. I have a question from our viewers from
Chicago. Keep your questions down to less than words, if you can.
I'm going to
abbreviate a little bit. What happens when product testing is
done, someone signs off on its accessibility
in their report, and an individual with a disability turns out
not to be able to use the product? It goes on. Do we run the risk
of getting into a situation where third-party
testing reports conflict with the experience of users? And if
so, will that third-party -- sorry, report, become
a shield for agencies who had the responsibility for making certain
their procurements were of accessible products? Did that make
sense?
KENNETH SALAETS: Yes.
STEVE JACOBS: I was trying to read it, not listen to what I was
reading.
KENNETH SALAETS: I have no idea who submitted that. Third-party
testing, if you go to the access board's website frx and the frequently
asked questions, I believe it's A, and it talks about the question
of compliance. We hear the word compliance a lot, for -- I don't
think that is unusual for people to talk about whether or not
the products are complying or
vendors or manufacturers are complying. But it has been made clear
that it's the federal agencies themselves that are obligated to
comply with section . The
amendments essentially put teeth into the law. has been in the
book or on the books whatever cliche you want to use, for many
years. But the
amendments enabled the federal employee, and in some case, private
citizens, to complain or protest or take stronger action relative
to the accessibility of products, simply
having a third-party test your products or a product that you're
contemplating buying, as a federal agency, does not remove, at
least according to our lawyers, the liability
that that agency has to still comply with . And hence it does
not necessarily preclude any type of legal or complaint process
being followed by an employer or private
citizen in certain cases cases.
JULIE CARROLL: Just one more question, Ken. If you've watched
the news recently, there has been some resistance on the part
of industry, not ITI, per se. But
some in industry, regarding third-party testing. And maybe you
can address what that is coming from, where that is coming from.
Is it
a fear of mandatory testing? Is it a fear that if some companies
do third-party testing, that then by default they all
will have to really in order to compete, and drive up the cost?
Can you speak to the concerns that we have been hearing about?
KENNETH SALAETS: I think there are numerous concerns. I don't
tend to sit around too often with individuals and discuss third-party
testing concerns. Frankly, we have tried to
move beyond this issue and our focus is primarily on the interoperability
and working closely with the assistive technology industry. Some
of the concerns might be the arbitrary or subjective
nature of third-party testing. It's not to say that there isn't
expertise in the marketplace, it's not to say that there aren't
resources out there that could be useful, for example, as a complement,
internal
testing or validating process. But we are opposed to mandatory
testing. We are opposed to agencies requiring it it. And
according to a copy of a letter I have to GSA administrator
Perry, representative Davis of Virginia is also opposed to third-party
testing, mandatory testing. A part of the concern is that it might
create a sense or
a misunderstanding of what that testing implies in terms of the
viability. I appreciate that Terri says that the TAG is not focusing
on certifications. I think initially there was
at least perhaps a misunderstanding but interpretation that that
was indeed, the case. And in talking to individuals like Dave
Drapkin of
the GSA acquisition policy shop and others, it was explicit that
there could be no certification or warranty requirements imposed
by agencies again, on manufacturers, or
resellers, ordealers, whatever the case might be, in terms of
who the government is -- from whom they are purchasing their products.
So, I think there is concern in that respect. Also, there is a
concern
that if you are developing new innovative approaches to accessibility,
addressing the needs of people with disabilities, that
if it's internal and particularly if it's proprietary and cutting
edge, let's say, using again that cliche, there is probably no
way that these outside
individuals or organizations would really have access to that
information, that approach, first of all, and so hence their evaluation
might be skewed
by not understanding the approach. Secondly, if it's a proprietary
approach that maybe is going to revolutionize a particular type
of way of addressing a particular disability, the company developing
that technology would want to be able to hit the marketplace running,
and not have their competitors necessarily have access to that
information prior to introducing the product. So
there is competitive concern and constraints in that respect.
JULIE CARROLL: Thank you very much. Next we are going to hear
from Allen Densmore. He is from the American
foundation for the blind. AFB has really been a catalyst to advance
accessible technology for many
many years, going back to the inception of section and all the
way through implementation of section . And Allen has certainly
been involved allle
way. He is also the co-chair of a technology task force of the
consortium for citizens with disabilities. Allen, thanks for coming.
Allen: Thank you very much. And Ken I was glad to hear the bit
of history, that your organization is actually older than ours.
Particularly in terms of your involvement in technology. We
date back to but we were not involved in technology. But for anybody
who remembers, the and a third LP record, for
anybody who can even remember records, that was a bees of technology
that we partnered in the development of, which then led to the
technology that made available the first talking books.
So, I don't quite go back that far. But I do go back far enough
in the various iterations of section to make what I
hope will make sense for people. One of the things that I think
everyone needs to notice about the way section is written is
that it is written such that the procurements that are undertaken
in conformance with section have the look and feel of procurements
that are undertaken generally.
Now, that happened for a very important reason. Because that process
was very thoroughly vetted before it went back into the legislation
by the individuals
both in the Office of management and budget, and in what are referred
to as the cognizant agencies, such that it would look exactly
like what they are
used to. And I think that's important. Because that's one of the
things that drives policy in this. Because the policy, as you've
heard, can occasionally
get very complicated. Procurement is sometimes verging on procurement
theology. But that's because procurement goes in that direction
generally. And
when you hear us involving ourselves in some of the questions
about how to make certain determinations as to whether something
is accessible, what you're listening to is
a debate that has the same look and feel of procurements generally.
The requiring officers are used to going through this. The procuring
officers are used to going
through this. And that's one of the things that driving . It also
drives us to try to solve some of these problems, or at least
to present what we think are reasonable
alternatives for solving some of these problems, so that a procurement
officer can make a good decision. From our -- my point of view
I think is similar
to what a procurement or requirement officer is all about, and
that's essentially trust but vair fivemt it is the verification
issue that is -- verify. It
is the verification issue that is very important. There is a high
level of trust. But often when you look at something as complex
as information technology, you have to be able to verify that
certain things
happened or should happen, so that you can make a decision about
whether what you're buying is in conformance with section . Now,
there are a number of way
that is that can happen. But again, what we feel very strongly
about is that companies vendor, should also be participating in
something that has, again,
that same look and feel. There may be complicated issues there,
but that procurement process is one that is understood. We
think companies understand fully in a lot of areas what they have
to deliver. And they understand, also, fully, or to the extent
that anyone can ever understand what's in the
mind, sometimes, of a requirement officer, that there are certain
benchmarks for the delivery of that product. Similar to what we
do with section . What
we would like to see is a situation, and I think there is a great
potential for this continuing, because the basis for this happening
is forming right now, such that companies
compete on the value of accessibility, and the delivery of accessibility.
Now, it's the value of
accessibility that I don't think any of us really debate about.
In a word, and you look at section , one of the things that you
need to remember is that it says that a federal employee with
a disability
or a customer of federal services with disability should be able
to acquire that information and use that information in the same
manner and time as
an individual without a disability. Now, that is a fairly simple
process to think about, but it's also why we got involved in trying
to set up certain benchmarks for how you'd
make judgments about whether that was happening. And then, in
turn, how a consumer could make a complaint if they did not think
that it was happening. Sorx we went into the whole
process with industry and with the advocacy groups and also with
the federal ocean of trying to look at what the accessibility
standards would be be. One of the other things that I think
is
important to remember about section is that in my view it is an
innovation driven piece of legislation. Because if you look at
it, one of the things that
it does talk about is refreshment of those standards. We are very
concerned as well, that a one-size fits all concept of disability
accessibility is not going to work. It's not even going to work
for individuals who are blind or visually impaired. But we are
also concerned that in the process of setting up some sort of
evaluation scheme, that you have to look
very carefully at something, as Ken was saying, that does not
set the bar so low that people will say well, this is all I have
to do. And that's all I'm going to do. We don't
think that's going to happen. We don't think that's going to happen
for a couple of reasons. Number one, again, going back to the
idea of the look and feel of procurement, procurement
officers are in the process not of doing a once and for all buy
of electronic and information technology, they do this on a constant
basis. They have
to refresh these requirements themselves. In turn, section very
clearly calls for a look back periodically to see what
does have to be refreshed in those standards. So we think that
the possibility that we are going to be stuck at a certain level
is probably not going to
happen, simply because of what drives procurement officers and
what the law says, in fact. So where do we go with this? Well,
as I said, we are very much from the point of view of
trust and verify. We think that the VPAT is an excellent way of
starting out. We have seen some really fine presentations through
that VPAT that, from our point of view, would allow a procurement
officer to look at something and say this is what they did. This
is how they did it. And this is how they say it works. And that's
really the
most that can be expected from something like this. There is,
I think, a problem in addition, though, in that when you do look
at this, there are also situations where some companies have not
done that.
Some companies have simply directed you to their home page that
talks about their commitment to accessibility. Well, that's an
easy one and most companies who compete
on that would probably look at their associate company and say
what are you doing? In competition, I would not think that would
work. But going back to
what we look at and what we look for is a process that's going
to give some benchmarks, which is why I think we like the concept
of independent
testing. And notice that I say that this is a concept. Independent
testing, from our point of view, is not a product evaluation system
that puts
a signed, sealed, approval on something. It is a process, though,
that tells you that when we looked at this product, we had certain
benchmarks about how that product should
perform. And that when we say benchmarks, they are something that
the company can look at and that the requirement officer or the
procurement officer
can look at as well to say this is the process that they went
through. Now, it's also possible that that process isn't going
to work work. And, in fact, we think, again, that this is
a reflection of the way the procurement process in general, works,
because there are sections of procurement, generally, where, for
example, they will test
software for compatibility with the basic operations of whatever
the application is supposed to be providing. And that's done independently.
And, in that process, there are
also situations where the particular device or software that's
being tested is probably not going to pass. But the fact of the
matter is, it also shows you what
you need to do. It gives a procurement officer some basis for
some objective judgment, because there has got to be a basis for
that in this process, such
that they can see that there were a certain set of processes that
went through, and that as near as we can get to something as complicated
in standard setting,
that these processes are roughly comparable. Now, that is something
that is going to be a moving target, over time. That is something
that we are all going
to learn from. I think the companies as well as the testing process
itself. But again, we think that something like that is substantially
in conformance with the way
procurement goes. It has the look and feel. I think another thing
that we would look for in a process is the involvement of qualified
individuals with disabilities.
This is something that we have seen as a practical example, wherein
there are certain claims that are made for accessibility that,
in
fact, if you looked at what the assumptions that the company was
making about its product, are probably accurate, as far as that
company knew
in terms of the applicability of that product to the general population.
What you do need to be able to do is to have individuals who are
qualified in
that process, to begin to look also in the verification scheme
of things to say: Not so much that we are the people who are going
to verify that this
is good, right and true, but that we are the people who because
of our background can also give you some indication of what you
ought to be looking for in terms
of the workability and the accessibility of that product. I think
in sum, we are very excited about this. We think that this is
a process
that can work and we think that the commitment that has been shown
by almost everyone that we have dealt with is very, very substantial,
primarily because they
think that it can deliver a range of exciting products, and that's
what we're looking for in this. We're not looking for a complaint
driven process. We are not looking
for a legally driven process. What we're looking for is something
that, again, has the look and feel that everybody is used to,
but begins to deliver products so
that people will have the access to those products and to the
information that was contemplated by the law law.
JULIE CARROLL: Thank you, Allen. I just have one quick question
for you. We know that some government agencies are performing
their own testing, rather
extensive in-house testing, before they buy a product. And then
we have organizations here who will do it for you or teach you
to do it in-house. Those
-- does the disability community that you know of have a position
as to who should be doing that testing or do they have a preference?
Or does it matter?
Allen: Well, I think it does matter. And substantially what the
technology and telecommunications task force of CCD has said is
that that should be independent testing. Now, with
respect to whether that resides outside of the government or whether
it resides in another entity, when we look at the concept of independent
testing, what we do
look for is the same benchmarks. If the federal government, if
the federal agency were to decide to do this in-house, that those
same benchmarks
exist that an independent unit would use.
JULIE CARROLL: Thank you. I'm afraid we are going to have to close
out this portion of our program. I want to thank our panelists.
Our audience really wants to get in on this. So we need
to move along. While the questions are coming across and Steve
is getting them organized, I do want to give you a quick overview
of ITTATC, the host of today's Web
cast. It stands for the information technology technical assistance
and training center. We are a project funded by NID DR, national
institute on disability and rehabilitation research. And we have
a number of partner, although the project is based at Georgia
Tech in Atlanta. Also involved is the world institute on
disabilities. The disability business technical assistance centers.
The information technology association of America. And the
University of Iowa's College of law, health policy and disability
center. Our mission is to advance the implementation of section
and
section for proceed -- through providing technical assistance
and training. And our primary stakeholders are industry, state
governments, and
consumers. And we encourage you to visit our website and use the
contact link there, and will the us know how we might be able
to assist you. Steve,
do we have some questions from the audience?
STEVE JACOBS: We do. And I should probably apologize. I know that
we are not going to get to everyone's question. So apologies ahead
of time. We have one question, how can I get a transcript
of this broadcast? I don't know.
JULIE CARROLL: I can answer that. There is a transcript of all
of our webcasts, posted after the event. It takes about hours
ordinarily to get it up up. And so it will be right here
at
TVWorldwide.com's website.
STEVE JACOBS: Okay. And we have several questions here, and I
thought I would ask Allen one question, and then maybe let the
panel see if they have any questions for each other.
Let's see. I remember the question, but it's in this stack and
I'm not finding it. It's fairly well-known that
does not focus on or address cognitive disabilities. So I was
going to ask Allen how he thinks that
issue should be addressed through any kind of testing. What are
your thoughts about persons can cognitive
disabilities, since the regulation didn't isolate certain groups,
it's talking about anybody accessing ENIT just
as anyone else can.
Allen: Well, there is no one size fits all for cognitive disabilities
much as in the same situation as other disabilities. What I think
developers should be looking at is that there are certain common
elements of accessibility of information. The examples that I
would give is that for an individual with cognitive
disabilities, there may be much the same as an individual with
a learning disability, or someone with low vision, as an example.
And that is what's on the screen? How
fast does it go by me on that screen? Do I have an ability to
shrink the information on that screen or to stap the
flow of the information on the screen so that I can read it? That's
an example, I think, of where you can cover some of these situations,
so that you don't have to
tease out each disability. And I think as we are moving along
we probably will learn more how to make some of those applications
across disability.
STEVE JACOBS: Thank you. I have a question from the audience for
Terri. And I don't have a city on this one. But I'll use it --
I do. It's Washington,
D.C.. I was going to use the person's name. It's from DC. Ms.
Youngblood's analogy about sending the paragraph to someone's
editor is interesting. You know what I'm talking about.
TERRI YOUNGBLOOD: Yes.
STEVE JACOBS: Okay. Good. That type of evaluation is enhairntly
subjective and some expressed concern that evaluations also
involve inherent subjectivity. How do TAG's testing processes
account for this subjectivity?
TERRI YOUNGBLOOD: It's an interesting question and I -- the comparison
with the paragraph was simply to simplify the concept of -- for
those in the audience that
aren't technical for testing methodologies. But the testing can
-- in terms of subjective, whether the testing is subjective,
it should be and can be in some
cases. But with the strict methodology on testing products, feature
by feature, I think you can use both types of measures for that.
STEVE JACOBS: Thank you. Julie?
JULIE CARROLL: Anymore questions from the audience? We have more
time.
STEVE JACOBS: Yes, we do. And this is from Florida. No. Georgia.
Excuse me. Could you please tell us how we
can determine that a product is accessible? Well, who should we
ask this question to? I'll let you collect. -- select.
JULIE CARROLL: Let's start with Brad and Dennis.
W. BRADLEY FAIN: The question of accessibility is a complex one
because there are so many definitions of what accessibility really
is. I guess if you go back to the -- back to the core definition
or the
lay person definition is that a person with a given disability,
and you can define the disability categories to some extent. There
may be some overlap. And then there is also a special case
where you have multiple disabilities that you have to consider.
But, if you define that user group, then that user group has to
be able to use the technology that we're talking about. So, in
our scheme of things, in addition to the checklist evaluation,
actually demonstrate that go a user with a given functional capability
or functional
limitation, however you choose to describe it, are able to make
use of the product effectively in the fashion that it was intended
to be used in. Collecting
that data and demonstrating that is one step towards demonstrating
accessibility. It's not the complete answer. There are other issues
to be considered, as everybody knows. But at least
it's one step toward demonstrating accessibility.
STEVE JACOBS: Thank you. Any other panelists --.
Was the question, how do you define accessibility?
STEVE JACOBS: The question was, can you tell us how we can determine
if a product is accessible.
That would be a whole day's show. However, given -- narrowing
the question, how can you determine whether an application, let's
use a software application,
is accessible to an individual who is visually impaired? -- impaired
it would be a better question to answer to draw a conclusion for
someone. And,
and specifically, if an individual uses a piece of technology,
like a screen reader, a screen reader interacts with a piece of
software in a
way that it can -- it speaks what is on the screen to the individual.
So, in order for an individual to use a software application,
the interaction between the
assistive technology and the software application must be evident.
The user must be able to locate all the controls on
the screen. They must be able to follow instructions. They must
be able to complete all the tasks that are involved with using
a piece of software application. So it's a very
complex question to answer.
STEVE JACOBS: Ken?
KENNETH SALAETS: At the risk of over as many reply fiing, I'm
going to assume that -- over simplifying, I want to assume that
the question goes to the notion of buying a
product and how do you determine whether a product, one over another,
might be better. In terms of my own experience, I recently bought
a
digital cam remarks I wasn't looking for accessibility. But there
were features that I was interested in. I looked at the product
itself and the representations made by the manufacturer, in this
case it was going to best buy and
seeing what was available on the shelf that I could actually touch,
examine it, evaluate, in my own respect. After that, I did my
own research. I looked on the Internet. I looked for
Internet magazines. I looked in digital camera magazines and the
like and made an evaluation based on the information that was
provided there. I would assume that if an individual is
contemplating making an acquisition, that you would have access
to different resources. Not everybody has access to the Internet.
We just assume they do, since we do, sometimes. But I think there
is probably
ample information in the marketplace, and in particular if the
manufacturer had a product that had certain accessible features,
I would assume that they would make that information available,
because that would be
a particular marketing aspect that they would want the customer
to know in the event that that particular feature was important
to the customer. So I would look to the manufacturer, and then
perhaps look at other
resources and then make that determination.
STEVE JACOBS: Thank you, Ken. Julie?
JULIE CARROLL: We have time for more questions.
STEVE JACOBS: We have time for more questions. What I thought
we might do, we all realize that accessibility is important and
that's why we're all
here. Usability is another term that is used that may encompass
more than accessibility. Some will define accessibility as being
part of
the usability equation. Designing a product that's usable globally
seems to make sense, at laes for industry. How
do you think the -- at least for industry. How should a development
process be approached or an assessment process be approached to
include usability
on a global basis? For example, earlier Allen and I talking about
designing an interface that is accessible to persons who are blind.
Well, there
are people who can't read, who can also use an interface designed
to talk to a person. And how can this be
taken into consideration so that economies of scale come into
play when a company wishes to design usable and oh, by the way,
accessible products? Let me address this first to Dennis and then
we can just go to everyone else.
DENNIS FOLDS: Thanks, Steve. I believe that the accessibility
is largely an extension of usability to a more complete and diverse
user population. One could
argue that making a product accessible is simply making it usable
to individuals who are not normally considered as part of the
user population. But they should be.
Usability is not a property of the device itself. But it's a property
of a person using the device for some stated purpose, for some
intended purpose. I
think you could -- if you could construe accessibility in the
same way, accessibility is not an integral property of the device,
but it's a property of a user with some
impairment intending to use the device or attempting to use the
device for some purpose. So, as far as the achieving sort of a
universal design approach or the
economies of scale for very large or mass marketing, I think it's
first of all essential to conceive of the user coplation in the
broadest terms. One
can identify the ranges of capabilities that are present there.
And you may find some important characteristics of users that
may not come under the traditional view of accessibility. I know
that I believe that it was Ken that noticed the issues with his
bifocals. We don't normally think of people that are simply wearing
corrective lenses as being part
of a population that is covered by the term accessibility. But
that's because it's so common. If it were not for the relative
frequency of this, we wouldn't think
of these as special users, as we develop products, whether they
are in automobiles or airplanes or whatever, we generally consider
the fact that someone might be wearing
bifocals. That is common enough that it has to be considered.
So usability needs to extend to include all of the special characteristics
that a diverse community of users harx, address
what are the requirements so that people with those properties
can make use of the device for certain intended purposes. With
that in mind, there will be a much broader market
for a product and I think the marketplace itself will sort it
out from there. Young yes. I agree. I also have something to add
regarding that. You said accessibility
was an extension of usability. And I agree. But I think usability
is a result of accessibility. I flipped it around. In my experience
in dealing with industry and particularly
software and Web Page design accessibility issues often if a developer
does something to increase the accessibility of a
product, it results in increasinged usability. For example, --
increased usability. For example on a Web Page, you have graphic
pictures. And a screen reader needs to get to the information
about that graphic, because the screen readers interpret page
content based on textual information. So
if there is what we call an Alt attribute to a graphic, then it's
accessible to a person using assistive
technology. It's also accessible to an individual using a hand
held device that may access the Web. For example, a cell phone.
So, for not only -- often
developers want to do the right thing, but they say well, you
know, I don't know anyone with a disability in this arena or blind
folks. Well, just because you make it accessible
to a person who is blind, you can increase the accessibility for
all of us. All of us are potentially users of AT. One out of five
individuals in the United States have a disability. So, by
the government enforcing and industry taking such a positive reaction
to , it will, overall, increase the accessibility of products
as a whole.
STEVE JACOBS: Thank you. Allen?
Allen: Ken may not share this opinion, because I think that his
industry and members think of themselves as great innovator but
I
think you guys are some of the great recyclers of all times and
I mean that as a complement and I think it may have import for
the question that was asked.
I was thinking, for example, of how it came to be that the PDA,
personal digital assistant device, got to be integrated with a
cell phone.
Or whether that happened the other way around, I'm not sure. But
the fact of the matter is there was somebody who was looking at
these two devices and the utility of the two devices, who
finally divided -- decide, we have these two things, why don't
we put them together and see what happens. And they did. And they
may have a market winner in this. And I'm thinking, it's that
kind
of approach, I think, that often happens in the industry that
can have some import for you, what you're talking
about, as we begin to look at what various accessible features
there are and what has been done to design these features into
products, and catalog and list them throughout our products. The
product developers, I expect are probably going to flip through
some of the pages from time to time and say let's put and together
and come up with a
brand new product, number .
STEVE JACOBS: Thank you, Allen. Terri mentioned something before
that I wanted to just pick up on briefly, and that was the mention
of including all
texts, makes transcoding a Web Page into wireless application
protocol more accessible, using simplified English as it's described
in the Web accessibility guide lanes,
also has a very powerful effect, certainly it would be easier
to read for a person with cognitive disabilities. But when global
companies translate
a Web Page or a website using simplified English uses about half
the words, which can cut the cost of translation in half
because the cost is usually by word. And it also translates more
accurately. I just got a question, this was over the telephone.
No
name. No city and I can see why you didn't want to leave your
name. This is for Terri. We told you this wasn't going to be easy.
No No. We don't. We didn't. A
question by phone call. Ms. Youngblood, you can tell in the beginning,
he should call you Terri. Terri stated industry doesn't have knowledge
to make evaluation of accessibility
of products, and that industry doesn't know how to evaluate vendors.
What was this based on? A study? A survey? What?
TERRI YOUNGBLOOD: My answer is more of a clarification. I didn't
say or intend to say that industry didn't know how to. Because
there are a lot of
industry leaders, and I like to refer to them as the big boy,
that really have an understanding of accessible technology. They
have employees with disabilities.
They have knowledge. But there are a lot of smaller and middle
sized companies that don't have those resources. And so if I generalized
industry as a whole, not knowing,
for that I apologize. Because there are some that really understand
it. I still believe that there are companies that have the in-house
expertise also need to have
a third-party evaluation. Because things will go missed. I work
for a large -- I worked in the past for a large manufacturer of
technology, who does
internal testing. But also seeks my expertise on their accessible
systems to review what they had already done. And that's just
good business and
infrastructure practice in my opinion. So clarification, there
is industry that knows what they are doing. But there are smaller
and middle sized companies that don't and that's what TAG is here
to assist and to help them. Thank you.
STEVE JACOBS: Thank you. Allen and Terri both were describing
a third-party testing environment. Is there a difference in your
approaches?
Allen, are you talking -- you're talking about mandatory testing,
and Terri you're not? I'm addressing this to Allen. Can you clarify
what position it is you're taking or who's taking and what that
position is?
Allen: I think the way we look at it is the federal government
and the procuring officer and agency have the responsibility
to do this. Now, whether they want to do it using independent
testing with an outside organization or do it within their agency,
that is going to be
largely up to what the procuring agency wants to do and what they
can do by their own internal rules and regulations. What we do
say, though, is that
that process, if it is internal or external, whenever it resides,
should have certain benchmarks to it, in terms of what it looks
at in terms
of product accessibility. And that we think, also, that one of
those benchmarks should be the involvement of customers and federal
employees who are qualified as
well. I hope that answers the question.
STEVE JACOBS: It does. And I thank you.
I have one question. I think an important part of the section
standard is the equivalent facilitation mechanism and I wonder
how each of your
testing methodologies take into account alternatives that you
might not have thought of when you were conducting the test. I'd
like to hear from
Brad, Terri, Dennis and Ken.
It's an interesting concept in . I believe it was put in to allow
for flexibility and perhaps innovating designs that were not particularly
considered when the
standards were developed. The equivalent facilitation approach,
if
a company, for example, wanted to come out and say that we have
this new design. We think it's very accessible. It doesn't directly
match up with the checklist items or the standards that
are in your checklist items items. So, how can we say that
it is accessible? Well, the only way, and I think the
access board would support that on the section government website,
is that the only way to test that is to evaluate it against the
functional performance criteria in subpart
C of section . And to me that suggests human performance testing
is required to demonstrate that, yes, this product
is accessible. Users with the types of disabilities outlined in
section can make use of this product in this alternate format
that perhaps doesn't directly
apply to the checklist items that are standards that section require.
So, I think the user and the
loop testing supports the demonstration of the equivalent of facilitation.
JULIE CARROLL: So that might be one area where the results of
the checklist might be different than the user in the lap testing !
testing.
It may be that the checklist item either fails or it is not applicable
to the item anymore, so that the procurement official really doesn't
have any basis for making
an accessibility judgment in absence of human performance data.
JULIE CARROLL: Terri, would you agree with that?
TERRI YOUNGBLOOD: Yes. I absolutely agree with it. I would like
to give a brief example of equivalent facilitation that I have
run into. I ran into a computer based
training module. CD, pop it in the computer. Up comes a training
manual and there is is a grafrk kal reputation of a training class
there.
That graphical representation is not, this particular one, was
not accessible assistive technology. For example, a screen reader
used by the blind
or visually impaired. The manufacturer had already -- this is
actually pre, but the story represents it. So the manufacturer
got to a certain point
and said well we put all this developmental work into it. In order
to make it check all the checklists that happened to be the proposed
standards at the time. But in order to meet all the technical
standards, we would have
to throw away the entire tool and start all over. So we said,
the manufacturer of this particular CBT said let's get creative.
And they looked to make the product usable,
if you do not have vision. They wanted it to be usable if you
do not have a certain level of mobility et cetera, et cetera.
And in fact, the result was they
incorporated voice files and keyboard accessibility. So, in fact,
if you went down the checklist on the software accessibility or
the software standards, it
couldn't meet them all. But if you looked at the functional performance
criteria of , is it usable by at least one mode, with a user without
vision, the answer was question. So, it
can be done for equivalent facilitation and I think it was put
there, the ak said board's intention was to put it there as technology
changes and grows,
room for innovation.
STEVE JACOBS: Any other comments from the panel?
I agree, exactly. That was the ak said board recognizing and accepting
some advice from industry, for that matter, that you couldn't
take
a single point in time and be able to create standards on accessibility
given the industry we are addressing and the types of products
and be able to provide a complete or a satisfactory
solution. So, I think we have viewed this as the window of opportunity
to did I know to advance our product designs and technology. And
the government was,
in its wisdom, in its innovative style, allowed that to happen.
I think it was helpful and very productive.
JULIE CARROLL: But how does the template figure in equivalent
facilitation? When really, you don't have to do any user testing
at all to complete the template.
What you would do, what I would assume, I never actually filled
out a template, is that you would take the particular section
of the standard that you're addressing, and identify
in your remarks how you addressed the needs of the individual
or the performance through a different means. It provides -- if
anything, it's an opportunity to showcase different approaches
to
addressing the same issue.
JULIE CARROLL: Thank you.
STEVE JACOBS: We received another question from the audience that
is very thought provoking. It's actually from Canada. And the
question is: Do any of the methodologies
attempt to go beyond section and address other countries' accessibility
issue, which raises an intriguing question. It looks like maybe
the
eyes of the world are looking this way to see what we're doing
and how we're doing it. And maybe we should just go very quickly
through the panel members to address that. Dennis?
-- I mean, Brad? Sorry.
W. BRADLEY FAIN: Putting my other hat on, I assisted it the --
in the needs assessment for the ITTATC project early on. And part
of that needs assessment was a
thorough literature review that I performed, both locally and
internationally, looking at the literature, seeing what guidance
was out there, trying to figure out what was important what wasn't
important.
Later, that served me very well in generating the checklist items
that I talked about earlier. So, as new
guidelines were available, to the extent that it was available
at the time that we did the literature review, which was about
a year ago, things are added to the database of checklist items,
if you will, I see it
as a continually evolving database. As soon as I learn about something
else that is accepted as a standard or as a guideline, I consider
it adding
-- I consider adding it to the checklist database that we used
in our accessibility evaluation.
STEVE JACOBS: Thank you, Dennis?
DENNIS FOLDS: Yes. I'll add to that. For a given evaluation, a
customer might be interested in purely section issues. However,
our methodology itself is
not inherently restricted to . In fact, it is inherently as broad
as a list of items can be. So
European and Canadian and other international standards become
available, as they become recognized and accepted, I
believe those will be added to our database of requirements. And
then it would be possible to fashion an evaluation in accordance
with some other specific standard that someone else might be interested
in.
STEVE JACOBS: Thank you. Terri?
TERRI YOUNGBLOOD: Yes. I agree with what he said. I know that
some other countries are looking to as a model guide for them.
And I think that's very crucial
in our thinking and laying out of methodology that has to include
the other countries and groups as well.
KENNETH SALAETS: One of the goals of Craig Lugard is to maintain
AT as quickly as possiblement and that was the intention of Congress
as well in addressing section .
So knowing that the federal government will have an billion dollar
market on an annual basis, the assumption is that
AT and IT manufacturers won't market slowlyly to the federal government.
And that's true. The intent is to build one product and be able
to sell it everywhere. ITI is committed
to not only addressing the section issue in the US, but also accessibility
issues worldwide. We are part of an organization called the international
information
infrastructure Congress, IIIC, that encompasses our sister associations,
technology associations throughout the world, and we anticipate
raising the issue of accessibility in that organization
and try to begin generating that dialog, so that there is Harmonization
between countries. Ideally, you ought to be able to take your
accessible laptop and be able
to use it anywhere you go, whether it's to access a wireless Web
or whatever the case may be. Ideally, it will reduce costs for
everybody
and it will expedite the introduction of new technology in the
marketplace. So we are happy to work with our European, Japanese,
Canadian friends, everyone,
we are hoping to have an international marketplace and approach
on this issue.
Allen: I think that Ken is right certainly with respect to information
technology. It is a global marketplace. What I would simply add
is that disability is a global issue, too. It
has varying impacts in varying countries. But one of the this
inks that we are noticing now with the worldwide net is that individuals
with disabilities are
signing on in greater numbers and discovering information in greater
numbers, and we hope that what we will be able to do is to get
them to go to places like by accessible, and begin
to look at what we are doing, indeed, as a model and also to see
how it can be adapted to their country and their culture.
STEVE JACOBS: Thank you. We just received another question from
the field. It seems to play on Ken's mention of the word mainstreaming
accessible
technology. And the question is as follows: To help convince product
vendors to design new products and change their existing products
so they are more accessible, it
would be helpful to have return on investment information. If
the ROI is strong, then it is a simple business decision to do
this extra work.
Do any of the panel members know of any accessibility return on
investment or cost benefit data they can share with product vendors?
Now, thank you for that very, very good question. Let's
just start back the other way, Allen?
Allen: I don't know of any specifically. But, I think that there
is a lot going on that is probably going to drive us in that direction.
Let me give you one example. When
the under thousand dollars computers came on the market, which
was probably, what, four or five years ago now at least, one of
the consulting groups that looks at marketing
took a look at who the newest purchasers of that, that is the
conquest purchaser, people who never owned a computer before.
Surprising statistic. The
well over half of them were individuals aged and older. Now, when
someone looks at that as a new conquest market,
I think that is important. What is embedded in that, at age ,
you're getting close to the time when with a number of factors
that are involved, you're probably not going to be able to see
that screen so
well after about four or five years of the use of your computer.
So I think that is going to drive some of this now.
STEVE JACOBS: Ken?
KENNETH SALAETS: I don't have any specific data but in terms of
responding to the federal marketplace, when you look at a billion
dollar market, I don't know if that is a cost benefit analysis,
but if you
are in the business of selling technology and you want to compete,
then you have to be responsive on this issue and you make your
determination.
TERRI YOUNGBLOOD: I don't have anything to add. I'm not familiar
with data.
DENNIS FOLDS: I'm familiar with qualitative analysis of opportunity
costs if one fails to make this investment, what potential market
is being lost. The data I'm more familiar
with is the projected aging population numbers. As was mentioned
earlier, properties of the aging population match up
well with properties of impairments that are covered by section
. And also including cognitive impairments as well. So if one
looks at the projections and
the size of the aging market in the United States in particular
over the next years, the opportunity costs of not being able to
serve that market might be profound.
STEVE JACOBS: Thank you. And Brad?
W. BRADLEY FAIN: I of course agree with Dennis. And I believe
that in performanceing the needs assessment
for ITTATC one of the most common things I'm asked is for that
type of data. Do you have any hard numbers? And I truly believe
they
don't exist sufficiently now. I've heard of several things, several
people that have been trying to develop this number -- those numbers,
but they don't have anything concrete to offer.
STEVE JACOBS: Julie in.
JULIE CARROLL: Well, I want to thank all of our panelists. We
are just about out of time. I'll give you one last chance if you
want to make a closing statement.
TERRI YOUNGBLOOD: I'd like to clarify something before a closing
statement. Earlier, I can't remember who mentioned it.
But they referred to TAG and Georgia Tech as competing methodologies.
And I wanted to clarify that that is not at all the case. TAG
is actually very interested in working
directly with Georgia Tech with the methodologies for testing
and hoping to develop a tight relationship between the two organizations.
STEVE JACOBS: Thank you.
JULIE CARROLL: Anyone else have any closing remarks? Young I have
one more thing. I think initially, when TAG was announced,
there was a misconception, one, that it was mandatory. And, two,
that it was a certification. And I think industry spoke strongly
against that,
which I can understand. So I want to clarify now that I think
that now that TAG is an incredible resource for government and
industry that doesn't have the
skill set to do it, and it is voluntary and it's not a certain
if I remember indication. It's there to assist the procurement
officials and market data and it's there
to assist industry in how to make their products more usable,
and that's our goal in the spirit of . And I wanted to clarify
that and ask you a question, based on those clarifications, do
you still feel industry
is against voluntary testing?
I don't think we ever said we are against voluntary testing. If
a company believes it would be a benefit of to invite a third-party
to evaluate, to
validate their own internal testing, that's what the marketplace
is about. It's an option. I just think everyone needs to understand
that, again, as far, if I'm a procurement
official and I'm looking for perhaps an added assurance, maybe
protection in some respect from the obligations, responsibilities
of section , I think there needs to be an understanding of
what it provides and doesn't provide. As far as manufacturers
are concerned, I believe and I know from my own companies, my
own experiences, that we
have highly qualified engineers. We do utilize people with disabilities
both as our employees and in other organizations to evaluate and
test our products. Think Steve can
probably address that more specifically than I could, in terms
of NCR's own experience experience. We are not operating
in a vacuum. If we try to develop technology in a vacuum, it would
sit on that shelf with those vacuum cleaners that don't pick up
the dust. It's a dynamic marketplace. You compete or die, and
to
be competitive you have to stay at least on if not ahead of the
curve. And on section we are committed to maintaining that process.
TERRI YOUNGBLOOD: I agree. I just wanted to clarify that. Thank
you.
JULIE CARROLL: Well, again, I thank all of our panelists. It's
been a very enlightening and thoughtful discussion. And I want
to thank the audience for their participation and some great questions.
I hope you will please
take a moment to look on your screen and complete the evaluation
form that is there. Let us know what we can do better next time
around. We will be having more webcasts. I would
also like to invite anyone out there who has methodologies or
tools that they would like to share, please contact us on the
contact link on our website, and
we would be happy to link to those, just to make sure resources
and -- make more resources and possibilities available to everyone
dealing with these
issues.
STEVE JACOBS: No closing remarks, except to thank Brad, Dennis,
Terri, Ken and Allen for sharing your thoughts, opinions, and
ideas with us. And our audience. It's been a pleasure
being here with you. Julie, thank you. We will have to do this
again.
JULIE CARROLL: Thank you. This concludes this webcast on product
accessibility evaluation methodologies. There will be future
webcasts and other types of technical assistance. You can do that
by visiting our website, which is www.ittatc.org.
|